mirror of
https://salsa.debian.org/ha-team/libqb
synced 2026-01-16 06:05:08 +00:00
The ringbuffer protocol uses the chunk magic number to indicate to the other side what state a chunk is in. It's therefore important to use strongly ordered memory writes to make sure that neither the compiler nor the CPU change the apparent order of the writes, since that would result in corrupted messages. Signed-off-by: Angus Salkeld <asalkeld@redhat.com>
868 lines
20 KiB
C
868 lines
20 KiB
C
/*
|
|
* Copyright (C) 2010-2011 Red Hat, Inc.
|
|
*
|
|
* Author: Angus Salkeld <asalkeld@redhat.com>
|
|
*
|
|
* This file is part of libqb.
|
|
*
|
|
* libqb is free software: you can redistribute it and/or modify
|
|
* it under the terms of the GNU Lesser General Public License as published by
|
|
* the Free Software Foundation, either version 2.1 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* libqb is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with libqb. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
#include "ringbuffer_int.h"
|
|
#include <qb/qbdefs.h>
|
|
#include <qb/qbatomic.h>
|
|
|
|
/*
|
|
* #define CRAZY_DEBUG_PRINTFS 1
|
|
*/
|
|
#ifdef CRAZY_DEBUG_PRINTFS
|
|
#define DEBUG_PRINTF(format, args...) \
|
|
do { \
|
|
printf(format, ##args); \
|
|
} while(0)
|
|
#else
|
|
#define DEBUG_PRINTF(format, args...)
|
|
#endif /* CRAZY_DEBUG_PRINTFS */
|
|
|
|
/*
|
|
* move the write pointer to the next 128 byte boundary
|
|
* write_pt goes in 4 bytes (sizeof(uint32_t))
|
|
* #define USE_CACHE_LINE_ALIGNMENT 1
|
|
*/
|
|
#ifdef USE_CACHE_LINE_ALIGNMENT
|
|
#define QB_CACHE_LINE_SIZE 128
|
|
#define QB_CACHE_LINE_WORDS (QB_CACHE_LINE_SIZE/sizeof(uint32_t))
|
|
#define idx_cache_line_step(idx) \
|
|
do { \
|
|
if (idx % QB_CACHE_LINE_WORDS) { \
|
|
idx += (QB_CACHE_LINE_WORDS - (idx % QB_CACHE_LINE_WORDS)); \
|
|
} \
|
|
if (idx > (rb->shared_hdr->word_size - 1)) { \
|
|
idx = ((idx) % (rb->shared_hdr->word_size)); \
|
|
} \
|
|
} while (0)
|
|
#else
|
|
#define QB_CACHE_LINE_SIZE 0
|
|
#define QB_CACHE_LINE_WORDS 0
|
|
#define idx_cache_line_step(idx) \
|
|
do { \
|
|
if (idx > (rb->shared_hdr->word_size - 1)) { \
|
|
idx = ((idx) % (rb->shared_hdr->word_size)); \
|
|
} \
|
|
} while (0)
|
|
#endif
|
|
|
|
|
|
/* the chunk header is two words
|
|
* 1) the chunk data size
|
|
* 2) the magic number
|
|
*/
|
|
#define QB_RB_CHUNK_HEADER_WORDS 2
|
|
#define QB_RB_CHUNK_HEADER_SIZE (sizeof(uint32_t) * QB_RB_CHUNK_HEADER_WORDS)
|
|
/*
|
|
* margin is the gap we leave when checking to see if we have enough
|
|
* space for a new chunk.
|
|
* So:
|
|
* qb_rb_space_free() >= QB_RB_CHUNK_MARGIN + new data chunk
|
|
* The extra word size is to allow for non word sized data chunks.
|
|
* QB_CACHE_LINE_WORDS is to make sure we have space to align the
|
|
* chunk.
|
|
*/
|
|
#define QB_RB_WORD_ALIGN 1
|
|
#define QB_RB_CHUNK_MARGIN (sizeof(uint32_t) * (QB_RB_CHUNK_HEADER_WORDS +\
|
|
QB_RB_WORD_ALIGN +\
|
|
QB_CACHE_LINE_WORDS))
|
|
#define QB_RB_CHUNK_MAGIC 0xA1A1A1A1
|
|
#define QB_RB_CHUNK_MAGIC_DEAD 0xD0D0D0D0
|
|
#define QB_RB_CHUNK_MAGIC_ALLOC 0xA110CED0
|
|
#define QB_RB_CHUNK_SIZE_GET(rb, pointer) \
|
|
qb_atomic_int_get(&rb->shared_data[pointer])
|
|
#define QB_RB_CHUNK_MAGIC_GET(rb, pointer) \
|
|
qb_atomic_int_get(&rb->shared_data[(pointer + 1) % rb->shared_hdr->word_size])
|
|
#define QB_RB_CHUNK_MAGIC_SET(rb, pointer, new_val) \
|
|
qb_atomic_int_set(&rb->shared_data[(pointer + 1) % rb->shared_hdr->word_size], new_val)
|
|
#define QB_RB_CHUNK_DATA_GET(rb, pointer) \
|
|
&rb->shared_data[(pointer + QB_RB_CHUNK_HEADER_WORDS) % rb->shared_hdr->word_size]
|
|
|
|
#define QB_MAGIC_ASSERT(_ptr_) \
|
|
do { \
|
|
uint32_t chunk_magic = QB_RB_CHUNK_MAGIC_GET(rb, _ptr_); \
|
|
if (chunk_magic != QB_RB_CHUNK_MAGIC) print_header(rb); \
|
|
assert(chunk_magic == QB_RB_CHUNK_MAGIC); \
|
|
} while (0)
|
|
|
|
#define idx_step(idx) \
|
|
do { \
|
|
if (idx > (rb->shared_hdr->word_size - 1)) { \
|
|
idx = ((idx) % (rb->shared_hdr->word_size)); \
|
|
} \
|
|
} while (0)
|
|
|
|
static void print_header(struct qb_ringbuffer_s * rb);
|
|
static void _rb_chunk_reclaim(struct qb_ringbuffer_s * rb);
|
|
|
|
qb_ringbuffer_t *
|
|
qb_rb_open(const char *name, size_t size, uint32_t flags,
|
|
size_t shared_user_data_size)
|
|
{
|
|
return qb_rb_open_2(name, size, flags, shared_user_data_size, NULL);
|
|
}
|
|
|
|
qb_ringbuffer_t *
|
|
qb_rb_open_2(const char *name, size_t size, uint32_t flags,
|
|
size_t shared_user_data_size,
|
|
struct qb_rb_notifier *notifiers)
|
|
{
|
|
struct qb_ringbuffer_s *rb;
|
|
size_t real_size;
|
|
size_t shared_size;
|
|
char path[PATH_MAX];
|
|
int32_t fd_hdr;
|
|
int32_t fd_data;
|
|
uint32_t file_flags = O_RDWR;
|
|
char filename[PATH_MAX];
|
|
int32_t error = 0;
|
|
void *shm_addr;
|
|
long page_size = sysconf(_SC_PAGESIZE);
|
|
|
|
#ifdef QB_FORCE_SHM_ALIGN
|
|
page_size = QB_MAX(page_size, 16 * 1024);
|
|
#endif /* QB_FORCE_SHM_ALIGN */
|
|
real_size = QB_ROUNDUP(size, page_size);
|
|
shared_size =
|
|
sizeof(struct qb_ringbuffer_shared_s) + shared_user_data_size;
|
|
|
|
if (flags & QB_RB_FLAG_CREATE) {
|
|
file_flags |= O_CREAT | O_TRUNC;
|
|
}
|
|
|
|
rb = calloc(1, sizeof(struct qb_ringbuffer_s));
|
|
if (rb == NULL) {
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* Create a shared_hdr memory segment for the header.
|
|
*/
|
|
snprintf(filename, PATH_MAX, "qb-%s-header", name);
|
|
fd_hdr = qb_sys_mmap_file_open(path, filename,
|
|
shared_size, file_flags);
|
|
if (fd_hdr < 0) {
|
|
error = fd_hdr;
|
|
qb_util_log(LOG_ERR, "couldn't create file for mmap");
|
|
goto cleanup_hdr;
|
|
}
|
|
|
|
rb->shared_hdr = mmap(0,
|
|
shared_size,
|
|
PROT_READ | PROT_WRITE, MAP_SHARED, fd_hdr, 0);
|
|
|
|
if (rb->shared_hdr == MAP_FAILED) {
|
|
error = -errno;
|
|
qb_util_log(LOG_ERR, "couldn't create mmap for header");
|
|
goto cleanup_hdr;
|
|
}
|
|
qb_atomic_init();
|
|
|
|
rb->flags = flags;
|
|
|
|
/*
|
|
* create the semaphore
|
|
*/
|
|
if (flags & QB_RB_FLAG_CREATE) {
|
|
rb->shared_data = NULL;
|
|
/* rb->shared_hdr->word_size tracks data by ints and not bytes/chars. */
|
|
rb->shared_hdr->word_size = real_size / sizeof(uint32_t);
|
|
rb->shared_hdr->write_pt = 0;
|
|
rb->shared_hdr->read_pt = 0;
|
|
(void)strlcpy(rb->shared_hdr->hdr_path, path, PATH_MAX);
|
|
}
|
|
if (notifiers && notifiers->post_fn) {
|
|
error = 0;
|
|
memcpy(&rb->notifier,
|
|
notifiers,
|
|
sizeof(struct qb_rb_notifier));
|
|
} else {
|
|
error = qb_rb_sem_create(rb, flags);
|
|
}
|
|
if (error < 0) {
|
|
errno = -error;
|
|
qb_util_perror(LOG_ERR, "couldn't create a semaphore");
|
|
goto cleanup_hdr;
|
|
}
|
|
|
|
/* Create the shared_data memory segment for the actual ringbuffer.
|
|
* They have to be separate.
|
|
*/
|
|
if (flags & QB_RB_FLAG_CREATE) {
|
|
snprintf(filename, PATH_MAX, "qb-%s-data", name);
|
|
fd_data = qb_sys_mmap_file_open(path,
|
|
filename,
|
|
real_size, file_flags);
|
|
(void)strlcpy(rb->shared_hdr->data_path, path, PATH_MAX);
|
|
} else {
|
|
fd_data = qb_sys_mmap_file_open(path,
|
|
rb->shared_hdr->data_path,
|
|
real_size, file_flags);
|
|
}
|
|
if (fd_data < 0) {
|
|
error = fd_data;
|
|
qb_util_log(LOG_ERR, "couldn't create file for mmap");
|
|
goto cleanup_hdr;
|
|
}
|
|
|
|
qb_util_log(LOG_DEBUG,
|
|
"shm size:%zd; real_size:%zd; rb->word_size:%d", size,
|
|
real_size, rb->shared_hdr->word_size);
|
|
|
|
error = qb_sys_circular_mmap(fd_data, &shm_addr, real_size);
|
|
rb->shared_data = shm_addr;
|
|
if (error != 0) {
|
|
qb_util_log(LOG_ERR, "couldn't create circular mmap on %s",
|
|
rb->shared_hdr->data_path);
|
|
goto cleanup_data;
|
|
}
|
|
|
|
if (flags & QB_RB_FLAG_CREATE) {
|
|
memset(rb->shared_data, 0, real_size);
|
|
rb->shared_data[rb->shared_hdr->word_size] = 5;
|
|
rb->shared_hdr->ref_count = 1;
|
|
} else {
|
|
qb_atomic_int_inc(&rb->shared_hdr->ref_count);
|
|
}
|
|
|
|
close(fd_hdr);
|
|
close(fd_data);
|
|
return rb;
|
|
|
|
cleanup_data:
|
|
close(fd_data);
|
|
if (flags & QB_RB_FLAG_CREATE) {
|
|
unlink(rb->shared_hdr->data_path);
|
|
}
|
|
|
|
cleanup_hdr:
|
|
if (fd_hdr >= 0) {
|
|
close(fd_hdr);
|
|
}
|
|
if (rb && (flags & QB_RB_FLAG_CREATE)) {
|
|
unlink(rb->shared_hdr->hdr_path);
|
|
if (rb->notifier.destroy_fn) {
|
|
(void)rb->notifier.destroy_fn(rb->notifier.instance);
|
|
}
|
|
}
|
|
if (rb && (rb->shared_hdr != MAP_FAILED && rb->shared_hdr != NULL)) {
|
|
munmap(rb->shared_hdr, sizeof(struct qb_ringbuffer_shared_s));
|
|
}
|
|
free(rb);
|
|
errno = -error;
|
|
return NULL;
|
|
}
|
|
|
|
|
|
void
|
|
qb_rb_close(struct qb_ringbuffer_s * rb)
|
|
{
|
|
if (rb == NULL) {
|
|
return;
|
|
}
|
|
qb_enter();
|
|
|
|
(void)qb_atomic_int_dec_and_test(&rb->shared_hdr->ref_count);
|
|
if (rb->flags & QB_RB_FLAG_CREATE) {
|
|
if (rb->notifier.destroy_fn) {
|
|
(void)rb->notifier.destroy_fn(rb->notifier.instance);
|
|
}
|
|
unlink(rb->shared_hdr->data_path);
|
|
unlink(rb->shared_hdr->hdr_path);
|
|
qb_util_log(LOG_DEBUG,
|
|
"Free'ing ringbuffer: %s",
|
|
rb->shared_hdr->hdr_path);
|
|
} else {
|
|
qb_util_log(LOG_DEBUG,
|
|
"Closing ringbuffer: %s", rb->shared_hdr->hdr_path);
|
|
}
|
|
munmap(rb->shared_data, (rb->shared_hdr->word_size * sizeof(uint32_t)) << 1);
|
|
munmap(rb->shared_hdr, sizeof(struct qb_ringbuffer_shared_s));
|
|
free(rb);
|
|
}
|
|
|
|
void
|
|
qb_rb_force_close(struct qb_ringbuffer_s * rb)
|
|
{
|
|
if (rb == NULL) {
|
|
return;
|
|
}
|
|
qb_enter();
|
|
|
|
if (rb->notifier.destroy_fn) {
|
|
(void)rb->notifier.destroy_fn(rb->notifier.instance);
|
|
}
|
|
|
|
errno = 0;
|
|
unlink(rb->shared_hdr->data_path);
|
|
qb_util_perror(LOG_DEBUG,
|
|
"Force free'ing ringbuffer: %s",
|
|
rb->shared_hdr->data_path);
|
|
|
|
errno = 0;
|
|
unlink(rb->shared_hdr->hdr_path);
|
|
qb_util_perror(LOG_DEBUG,
|
|
"Force free'ing ringbuffer: %s",
|
|
rb->shared_hdr->hdr_path);
|
|
munmap(rb->shared_data, (rb->shared_hdr->word_size * sizeof(uint32_t)) << 1);
|
|
munmap(rb->shared_hdr, sizeof(struct qb_ringbuffer_shared_s));
|
|
free(rb);
|
|
}
|
|
|
|
char *
|
|
qb_rb_name_get(struct qb_ringbuffer_s * rb)
|
|
{
|
|
if (rb == NULL) {
|
|
return NULL;
|
|
}
|
|
return rb->shared_hdr->hdr_path;
|
|
}
|
|
|
|
void *
|
|
qb_rb_shared_user_data_get(struct qb_ringbuffer_s * rb)
|
|
{
|
|
if (rb == NULL) {
|
|
return NULL;
|
|
}
|
|
return rb->shared_hdr->user_data;
|
|
}
|
|
|
|
int32_t
|
|
qb_rb_refcount_get(struct qb_ringbuffer_s * rb)
|
|
{
|
|
if (rb == NULL) {
|
|
return -EINVAL;
|
|
}
|
|
return qb_atomic_int_get(&rb->shared_hdr->ref_count);
|
|
}
|
|
|
|
ssize_t
|
|
qb_rb_space_free(struct qb_ringbuffer_s * rb)
|
|
{
|
|
uint32_t write_size;
|
|
uint32_t read_size;
|
|
size_t space_free = 0;
|
|
|
|
if (rb == NULL) {
|
|
return -EINVAL;
|
|
}
|
|
if (rb->notifier.space_used_fn) {
|
|
return (rb->shared_hdr->word_size * sizeof(uint32_t)) -
|
|
rb->notifier.space_used_fn(rb->notifier.instance);
|
|
}
|
|
write_size = rb->shared_hdr->write_pt;
|
|
read_size = rb->shared_hdr->read_pt;
|
|
|
|
if (write_size > read_size) {
|
|
space_free =
|
|
(read_size - write_size + rb->shared_hdr->word_size) - 1;
|
|
} else if (write_size < read_size) {
|
|
space_free = (read_size - write_size) - 1;
|
|
} else {
|
|
if (rb->notifier.q_len_fn && rb->notifier.q_len_fn(rb->notifier.instance) > 0) {
|
|
space_free = 0;
|
|
} else {
|
|
space_free = rb->shared_hdr->word_size;
|
|
}
|
|
}
|
|
|
|
/* word -> bytes */
|
|
return (space_free * sizeof(uint32_t));
|
|
}
|
|
|
|
ssize_t
|
|
qb_rb_space_used(struct qb_ringbuffer_s * rb)
|
|
{
|
|
uint32_t write_size;
|
|
uint32_t read_size;
|
|
size_t space_used;
|
|
|
|
if (rb == NULL) {
|
|
return -EINVAL;
|
|
}
|
|
if (rb->notifier.space_used_fn) {
|
|
return rb->notifier.space_used_fn(rb->notifier.instance);
|
|
}
|
|
write_size = rb->shared_hdr->write_pt;
|
|
read_size = rb->shared_hdr->read_pt;
|
|
|
|
if (write_size > read_size) {
|
|
space_used = write_size - read_size;
|
|
} else if (write_size < read_size) {
|
|
space_used =
|
|
(write_size - read_size + rb->shared_hdr->word_size) - 1;
|
|
} else {
|
|
space_used = 0;
|
|
}
|
|
/* word -> bytes */
|
|
return (space_used * sizeof(uint32_t));
|
|
}
|
|
|
|
ssize_t
|
|
qb_rb_chunks_used(struct qb_ringbuffer_s *rb)
|
|
{
|
|
if (rb == NULL) {
|
|
return -EINVAL;
|
|
}
|
|
if (rb->notifier.q_len_fn) {
|
|
return rb->notifier.q_len_fn(rb->notifier.instance);
|
|
}
|
|
return -ENOTSUP;
|
|
}
|
|
|
|
void *
|
|
qb_rb_chunk_alloc(struct qb_ringbuffer_s * rb, size_t len)
|
|
{
|
|
uint32_t write_pt;
|
|
|
|
if (rb == NULL) {
|
|
errno = EINVAL;
|
|
return NULL;
|
|
}
|
|
/*
|
|
* Reclaim data if we are over writing and we need space
|
|
*/
|
|
if (rb->flags & QB_RB_FLAG_OVERWRITE) {
|
|
while (qb_rb_space_free(rb) < (len + QB_RB_CHUNK_MARGIN)) {
|
|
_rb_chunk_reclaim(rb);
|
|
}
|
|
} else {
|
|
if (qb_rb_space_free(rb) < (len + QB_RB_CHUNK_MARGIN)) {
|
|
errno = EAGAIN;
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
write_pt = rb->shared_hdr->write_pt;
|
|
/*
|
|
* insert the chunk header
|
|
*/
|
|
rb->shared_data[write_pt] = 0;
|
|
QB_RB_CHUNK_MAGIC_SET(rb, write_pt, QB_RB_CHUNK_MAGIC_ALLOC);
|
|
|
|
/*
|
|
* return a pointer to the beginning of the chunk data
|
|
*/
|
|
return (void *)QB_RB_CHUNK_DATA_GET(rb, write_pt);
|
|
|
|
}
|
|
|
|
static uint32_t
|
|
qb_rb_chunk_step(struct qb_ringbuffer_s * rb, uint32_t pointer)
|
|
{
|
|
uint32_t chunk_size = QB_RB_CHUNK_SIZE_GET(rb, pointer);
|
|
/*
|
|
* skip over the chunk header
|
|
*/
|
|
pointer += QB_RB_CHUNK_HEADER_WORDS;
|
|
|
|
/*
|
|
* skip over the user's data.
|
|
*/
|
|
pointer += (chunk_size / sizeof(uint32_t));
|
|
/* make allowance for non-word sizes */
|
|
if ((chunk_size % (sizeof(uint32_t) * QB_RB_WORD_ALIGN)) != 0) {
|
|
pointer++;
|
|
}
|
|
|
|
idx_cache_line_step(pointer);
|
|
return pointer;
|
|
}
|
|
|
|
int32_t
|
|
qb_rb_chunk_commit(struct qb_ringbuffer_s * rb, size_t len)
|
|
{
|
|
uint32_t old_write_pt;
|
|
|
|
if (rb == NULL) {
|
|
return -EINVAL;
|
|
}
|
|
/*
|
|
* commit the magic & chunk_size
|
|
*/
|
|
old_write_pt = rb->shared_hdr->write_pt;
|
|
rb->shared_data[old_write_pt] = len;
|
|
|
|
/*
|
|
* commit the new write pointer
|
|
*/
|
|
rb->shared_hdr->write_pt = qb_rb_chunk_step(rb, old_write_pt);
|
|
QB_RB_CHUNK_MAGIC_SET(rb, old_write_pt, QB_RB_CHUNK_MAGIC);
|
|
|
|
DEBUG_PRINTF("commit [%zd] read: %u, write: %u -> %u (%u)\n",
|
|
(rb->notifier.q_len_fn ?
|
|
rb->notifier.q_len_fn(rb->notifier.instance) : 0),
|
|
rb->shared_hdr->read_pt,
|
|
old_write_pt,
|
|
rb->shared_hdr->write_pt,
|
|
rb->shared_hdr->word_size);
|
|
|
|
/*
|
|
* post the notification to the reader
|
|
*/
|
|
if (rb->notifier.post_fn) {
|
|
return rb->notifier.post_fn(rb->notifier.instance, len);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
ssize_t
|
|
qb_rb_chunk_write(struct qb_ringbuffer_s * rb, const void *data, size_t len)
|
|
{
|
|
char *dest = qb_rb_chunk_alloc(rb, len);
|
|
int32_t res = 0;
|
|
|
|
if (rb == NULL) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (dest == NULL) {
|
|
return -errno;
|
|
}
|
|
|
|
memcpy(dest, data, len);
|
|
|
|
res = qb_rb_chunk_commit(rb, len);
|
|
if (res < 0) {
|
|
return res;
|
|
}
|
|
|
|
return len;
|
|
}
|
|
|
|
static void
|
|
_rb_chunk_reclaim(struct qb_ringbuffer_s * rb)
|
|
{
|
|
uint32_t old_read_pt;
|
|
uint32_t new_read_pt;
|
|
uint32_t old_chunk_size;
|
|
uint32_t chunk_magic;
|
|
|
|
old_read_pt = rb->shared_hdr->read_pt;
|
|
chunk_magic = QB_RB_CHUNK_MAGIC_GET(rb, old_read_pt);
|
|
if (chunk_magic != QB_RB_CHUNK_MAGIC) {
|
|
return;
|
|
}
|
|
|
|
old_chunk_size = QB_RB_CHUNK_SIZE_GET(rb, old_read_pt);
|
|
new_read_pt = qb_rb_chunk_step(rb, old_read_pt);
|
|
|
|
/*
|
|
* clear the header
|
|
*/
|
|
rb->shared_data[old_read_pt] = 0;
|
|
QB_RB_CHUNK_MAGIC_SET(rb, old_read_pt, QB_RB_CHUNK_MAGIC_DEAD);
|
|
|
|
/*
|
|
* set the new read pointer after clearing the header
|
|
* to prevent a situation where a fast writer will write their
|
|
* new chunk between setting the new read pointer and clearing the
|
|
* header.
|
|
*/
|
|
rb->shared_hdr->read_pt = new_read_pt;
|
|
|
|
if (rb->notifier.reclaim_fn) {
|
|
int rc = rb->notifier.reclaim_fn(rb->notifier.instance,
|
|
old_chunk_size);
|
|
if (rc < 0) {
|
|
errno = -rc;
|
|
qb_util_perror(LOG_WARNING, "reclaim_fn");
|
|
}
|
|
}
|
|
|
|
DEBUG_PRINTF("reclaim [%zd]: read: %u -> %u, write: %u\n",
|
|
(rb->notifier.q_len_fn ?
|
|
rb->notifier.q_len_fn(rb->notifier.instance) : 0),
|
|
old_read_pt,
|
|
rb->shared_hdr->read_pt,
|
|
rb->shared_hdr->write_pt);
|
|
}
|
|
|
|
void
|
|
qb_rb_chunk_reclaim(struct qb_ringbuffer_s * rb)
|
|
{
|
|
if (rb == NULL) {
|
|
return;
|
|
}
|
|
_rb_chunk_reclaim(rb);
|
|
}
|
|
|
|
ssize_t
|
|
qb_rb_chunk_peek(struct qb_ringbuffer_s * rb, void **data_out, int32_t timeout)
|
|
{
|
|
uint32_t read_pt;
|
|
uint32_t chunk_size;
|
|
uint32_t chunk_magic;
|
|
int32_t res = 0;
|
|
|
|
if (rb == NULL) {
|
|
return -EINVAL;
|
|
}
|
|
if (rb->notifier.timedwait_fn) {
|
|
res = rb->notifier.timedwait_fn(rb->notifier.instance, timeout);
|
|
}
|
|
if (res < 0 && res != -EIDRM) {
|
|
if (res == -ETIMEDOUT) {
|
|
return 0;
|
|
} else {
|
|
errno = -res;
|
|
qb_util_perror(LOG_ERR, "sem_timedwait");
|
|
}
|
|
return res;
|
|
}
|
|
read_pt = rb->shared_hdr->read_pt;
|
|
chunk_magic = QB_RB_CHUNK_MAGIC_GET(rb, read_pt);
|
|
if (chunk_magic != QB_RB_CHUNK_MAGIC) {
|
|
if (rb->notifier.post_fn) {
|
|
(void)rb->notifier.post_fn(rb->notifier.instance, res);
|
|
}
|
|
return 0;
|
|
}
|
|
chunk_size = QB_RB_CHUNK_SIZE_GET(rb, read_pt);
|
|
*data_out = QB_RB_CHUNK_DATA_GET(rb, read_pt);
|
|
return chunk_size;
|
|
}
|
|
|
|
ssize_t
|
|
qb_rb_chunk_read(struct qb_ringbuffer_s * rb, void *data_out, size_t len,
|
|
int32_t timeout)
|
|
{
|
|
uint32_t read_pt;
|
|
uint32_t chunk_size;
|
|
uint32_t chunk_magic;
|
|
int32_t res = 0;
|
|
|
|
if (rb == NULL) {
|
|
return -EINVAL;
|
|
}
|
|
if (rb->notifier.timedwait_fn) {
|
|
res = rb->notifier.timedwait_fn(rb->notifier.instance, timeout);
|
|
}
|
|
if (res < 0 && res != -EIDRM) {
|
|
if (res != -ETIMEDOUT) {
|
|
errno = -res;
|
|
qb_util_perror(LOG_ERR, "sem_timedwait");
|
|
}
|
|
return res;
|
|
}
|
|
|
|
read_pt = rb->shared_hdr->read_pt;
|
|
chunk_magic = QB_RB_CHUNK_MAGIC_GET(rb, read_pt);
|
|
|
|
if (chunk_magic != QB_RB_CHUNK_MAGIC) {
|
|
if (rb->notifier.timedwait_fn == NULL) {
|
|
return -ETIMEDOUT;
|
|
} else {
|
|
(void)rb->notifier.post_fn(rb->notifier.instance, res);
|
|
#ifdef EBADMSG
|
|
return -EBADMSG;
|
|
#else
|
|
return -EINVAL;
|
|
#endif
|
|
}
|
|
}
|
|
|
|
chunk_size = QB_RB_CHUNK_SIZE_GET(rb, read_pt);
|
|
if (len < chunk_size) {
|
|
qb_util_log(LOG_ERR,
|
|
"trying to recv chunk of size %d but %d available",
|
|
len, chunk_size);
|
|
if (rb->notifier.post_fn) {
|
|
(void)rb->notifier.post_fn(rb->notifier.instance, chunk_size);
|
|
}
|
|
return -ENOBUFS;
|
|
}
|
|
|
|
memcpy(data_out,
|
|
QB_RB_CHUNK_DATA_GET(rb, read_pt),
|
|
chunk_size);
|
|
|
|
_rb_chunk_reclaim(rb);
|
|
|
|
return chunk_size;
|
|
}
|
|
|
|
static void
|
|
print_header(struct qb_ringbuffer_s * rb)
|
|
{
|
|
printf("Ringbuffer: \n");
|
|
if (rb->flags & QB_RB_FLAG_OVERWRITE) {
|
|
printf(" ->OVERWRITE\n");
|
|
} else {
|
|
printf(" ->NORMAL\n");
|
|
}
|
|
printf(" ->write_pt [%d]\n", rb->shared_hdr->write_pt);
|
|
printf(" ->read_pt [%d]\n", rb->shared_hdr->read_pt);
|
|
printf(" ->size [%d words]\n", rb->shared_hdr->word_size);
|
|
#ifndef S_SPLINT_S
|
|
printf(" =>free [%zu bytes]\n", qb_rb_space_free(rb));
|
|
printf(" =>used [%zu bytes]\n", qb_rb_space_used(rb));
|
|
#endif /* S_SPLINT_S */
|
|
}
|
|
|
|
ssize_t
|
|
qb_rb_write_to_file(struct qb_ringbuffer_s * rb, int32_t fd)
|
|
{
|
|
ssize_t result;
|
|
ssize_t written_size = 0;
|
|
|
|
if (rb == NULL) {
|
|
return -EINVAL;
|
|
}
|
|
print_header(rb);
|
|
|
|
result = write(fd, &rb->shared_hdr->word_size, sizeof(uint32_t));
|
|
if (result != sizeof(uint32_t)) {
|
|
return -errno;
|
|
}
|
|
written_size += result;
|
|
|
|
/*
|
|
* store the read & write pointers
|
|
*/
|
|
result = write(fd, (void *)&rb->shared_hdr->write_pt, sizeof(uint32_t));
|
|
if (result != sizeof(uint32_t)) {
|
|
return -errno;
|
|
}
|
|
written_size += result;
|
|
result = write(fd, (void *)&rb->shared_hdr->read_pt, sizeof(uint32_t));
|
|
if (result != sizeof(uint32_t)) {
|
|
return -errno;
|
|
}
|
|
written_size += result;
|
|
|
|
result = write(fd, rb->shared_data,
|
|
rb->shared_hdr->word_size * sizeof(uint32_t));
|
|
if (result != rb->shared_hdr->word_size * sizeof(uint32_t)) {
|
|
return -errno;
|
|
}
|
|
written_size += result;
|
|
|
|
qb_util_log(LOG_DEBUG, " writing total of: %zd\n", written_size);
|
|
|
|
return written_size;
|
|
}
|
|
|
|
qb_ringbuffer_t *
|
|
qb_rb_create_from_file(int32_t fd, uint32_t flags)
|
|
{
|
|
ssize_t n_read;
|
|
size_t n_required;
|
|
size_t total_read = 0;
|
|
uint32_t read_pt;
|
|
uint32_t write_pt;
|
|
struct qb_ringbuffer_s *rb;
|
|
uint32_t word_size = 0;
|
|
|
|
if (fd < 0) {
|
|
return NULL;
|
|
}
|
|
|
|
n_required = sizeof(uint32_t);
|
|
n_read = read(fd, &word_size, n_required);
|
|
if (n_read != n_required) {
|
|
qb_util_perror(LOG_ERR, "Unable to read blackbox file header");
|
|
return NULL;
|
|
}
|
|
total_read += n_read;
|
|
|
|
n_read = read(fd, &write_pt, sizeof(uint32_t));
|
|
assert(n_read == sizeof(uint32_t));
|
|
total_read += n_read;
|
|
|
|
n_read = read(fd, &read_pt, sizeof(uint32_t));
|
|
assert(n_read == sizeof(uint32_t));
|
|
total_read += n_read;
|
|
|
|
if (write_pt > word_size || read_pt > word_size) {
|
|
qb_util_log(LOG_ERR, "Corrupt blackbox: read_pt:%d, write_pt:%d, word_size:%d",
|
|
read_pt, write_pt, word_size);
|
|
return NULL;
|
|
}
|
|
|
|
n_required = (word_size * sizeof(uint32_t));
|
|
rb = qb_rb_open("create_from_file", n_required,
|
|
QB_RB_FLAG_CREATE | QB_RB_FLAG_NO_SEMAPHORE, 0);
|
|
if (rb == NULL) {
|
|
return NULL;
|
|
}
|
|
rb->shared_hdr->read_pt = read_pt;
|
|
rb->shared_hdr->write_pt = write_pt;
|
|
|
|
n_read = read(fd, rb->shared_data, n_required);
|
|
if (n_read < 0) {
|
|
qb_util_perror(LOG_ERR, "Unable to read blackbox file data");
|
|
goto cleanup_fail;
|
|
}
|
|
total_read += n_read;
|
|
|
|
if (n_read != n_required) {
|
|
qb_util_log(LOG_WARNING, "read %zd bytes, but expected %zu",
|
|
n_read, n_required);
|
|
goto cleanup_fail;
|
|
}
|
|
|
|
qb_util_log(LOG_DEBUG, "read total of: %zd", total_read);
|
|
print_header(rb);
|
|
|
|
return rb;
|
|
|
|
cleanup_fail:
|
|
qb_rb_close(rb);
|
|
return NULL;
|
|
}
|
|
|
|
int32_t
|
|
qb_rb_chown(struct qb_ringbuffer_s * rb, uid_t owner, gid_t group)
|
|
{
|
|
int32_t res;
|
|
|
|
if (rb == NULL) {
|
|
return -EINVAL;
|
|
}
|
|
res = chown(rb->shared_hdr->data_path, owner, group);
|
|
if (res < 0 && errno != EPERM) {
|
|
return -errno;
|
|
}
|
|
res = chown(rb->shared_hdr->hdr_path, owner, group);
|
|
if (res < 0 && errno != EPERM) {
|
|
return -errno;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int32_t
|
|
qb_rb_chmod(qb_ringbuffer_t * rb, mode_t mode)
|
|
{
|
|
int32_t res;
|
|
|
|
if (rb == NULL) {
|
|
return -EINVAL;
|
|
}
|
|
res = chmod(rb->shared_hdr->data_path, mode);
|
|
if (res < 0) {
|
|
return -errno;
|
|
}
|
|
res = chmod(rb->shared_hdr->hdr_path, mode);
|
|
if (res < 0) {
|
|
return -errno;
|
|
}
|
|
return 0;
|
|
}
|