/***************************************************************************** # # # uStreamer - Lightweight and fast MJPG-HTTP streamer. # # # # Copyright (C) 2018 Maxim Devaev # # # # This program is free software: you can redistribute it and/or modify # # it under the terms of the GNU General Public License as published by # # the Free Software Foundation, either version 3 of the License, or # # (at your option) any later version. # # # # This program is distributed in the hope that it will be useful, # # but WITHOUT ANY WARRANTY; without even the implied warranty of # # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # # GNU General Public License for more details. # # # # You should have received a copy of the GNU General Public License # # along with this program. If not, see . # # # *****************************************************************************/ #include #include #include #include #include #include #include #include #include "tools.h" #include "logging.h" #include "xioctl.h" #include "device.h" #include "encoder.h" #include "stream.h" #ifdef WITH_WORKERS_GPIO_DEBUG # include # ifndef WORKERS_GPIO_DEBUG_START_PIN # define WORKERS_GPIO_DEBUG_START_PIN 5 # endif #endif static bool _stream_wait_worker(struct stream_t *stream, struct workers_pool_t *pool, struct worker_t **ready_worker); static void _stream_expose_picture(struct stream_t *stream, unsigned buf_index); static void _stream_give_buf_to_worker(struct workers_pool_t *pool, struct worker_t *free_worker, unsigned buf_index); static long double _stream_get_fluency_delay(struct device_t *dev, struct workers_pool_t *pool); static int _stream_init_loop(struct stream_t *stream, struct workers_pool_t *pool); static int _stream_init(struct stream_t *stream, struct workers_pool_t *pool); static void _stream_init_workers(struct stream_t *stream, struct workers_pool_t *pool); static void *_stream_worker_thread(void *v_worker); static void _stream_destroy_workers(struct stream_t *stream, struct workers_pool_t *pool); struct stream_t *stream_init(struct device_t *dev, struct encoder_t *encoder) { struct process_t *proc; struct stream_t *stream; A_CALLOC(proc, 1); atomic_init(&proc->stop, false); atomic_init(&proc->slowdown, false); A_CALLOC(stream, 1); stream->dev = dev; stream->encoder = encoder; atomic_init(&stream->updated, false); A_MUTEX_INIT(&stream->mutex); stream->proc = proc; return stream; } void stream_destroy(struct stream_t *stream) { A_MUTEX_DESTROY(&stream->mutex); free(stream->proc); free(stream); } void stream_loop(struct stream_t *stream) { struct workers_pool_t pool; MEMSET_ZERO(pool); atomic_init(&pool.workers_stop, false); pool.encoder = stream->encoder; LOG_INFO("Using V4L2 device: %s", stream->dev->path); LOG_INFO("Using desired FPS: %u", stream->dev->desired_fps); while (_stream_init_loop(stream, &pool) == 0) { long double grab_after = 0; unsigned fluency_passed = 0; unsigned captured_fps_accum = 0; long long captured_fps_second = 0; bool persistent_timeout_reported = false; LOG_DEBUG("Allocation memory for stream picture ..."); A_CALLOC(stream->picture.data, stream->dev->run->max_raw_image_size); LOG_INFO("Capturing ..."); while (!atomic_load(&stream->proc->stop)) { struct worker_t *ready_worker; SEP_DEBUG('-'); LOG_DEBUG("Waiting for worker ..."); if (_stream_wait_worker(stream, &pool, &ready_worker)) { _stream_expose_picture(stream, ready_worker->buf_index); LOG_PERF("##### Encoded picture exposed; worker = %u", ready_worker->number); } else { LOG_PERF("----- Encoded picture dropped; worker = %u", ready_worker->number); } if (ready_worker == NULL) { break; } if (atomic_load(&stream->proc->stop)) { break; } if (atomic_load(&stream->proc->slowdown)) { usleep(1000000); } bool has_read; bool has_write; bool has_error; int selected = device_select(stream->dev, &has_read, &has_write, &has_error); if (selected < 0) { if (errno != EINTR) { LOG_PERROR("Mainloop select() error"); break; } } else if (selected == 0) { if (stream->dev->persistent) { if (!persistent_timeout_reported) { LOG_ERROR("Mainloop select() timeout, polling ...") persistent_timeout_reported = true; } continue; } else { LOG_ERROR("Mainloop select() timeout"); break; } } else { persistent_timeout_reported = false; if (has_read) { LOG_DEBUG("Frame is ready"); int buf_index; long double now = get_now_monotonic(); long long now_second = floor_ms(now); if ((buf_index = device_grab_buffer(stream->dev)) < 0) { break; } stream->dev->run->pictures[buf_index].grab_time = now; // Workaround for broken, corrupted frames: // Under low light conditions corrupted frames may get captured. // The good thing is such frames are quite small compared to the regular pictures. // For example a VGA (640x480) webcam picture is normally >= 8kByte large, // corrupted frames are smaller. if (stream->dev->run->hw_buffers[buf_index].used < stream->dev->min_frame_size) { LOG_DEBUG("Dropping too small frame sized %zu bytes, assuming it as broken", stream->dev->run->hw_buffers[buf_index].used); goto pass_frame; } { if (now < grab_after) { fluency_passed += 1; LOG_VERBOSE("Passed %u frames for fluency: now=%.03Lf; grab_after=%.03Lf", fluency_passed, now, grab_after); goto pass_frame; } fluency_passed = 0; if (now_second != captured_fps_second) { stream->captured_fps = captured_fps_accum; captured_fps_accum = 0; captured_fps_second = now_second; LOG_PERF("A new second has come, Captured-FPS = %u", stream->captured_fps); } captured_fps_accum += 1; long double fluency_delay = _stream_get_fluency_delay(stream->dev, &pool); grab_after = now + fluency_delay; LOG_VERBOSE("Fluency: delay=%.03Lf; grab_after=%.03Lf", fluency_delay, grab_after); } LOG_DEBUG("Grabbed a new frame to buffer %u", buf_index); _stream_give_buf_to_worker(&pool, ready_worker, buf_index); goto next_handlers; // Поток сам освободит буфер pass_frame: if (device_release_buffer(stream->dev, buf_index) < 0) { break; } } next_handlers: if (has_write) { LOG_ERROR("Got unexpected writing event, seems device was disconnected"); break; } if (has_error) { LOG_INFO("Got V4L2 event"); if (device_consume_event(stream->dev) < 0) { break; } } } } A_MUTEX_LOCK(&stream->mutex); stream->picture.used = 0; // On stream offline free(stream->picture.data); stream->width = 0; stream->height = 0; atomic_store(&stream->updated, true); A_MUTEX_UNLOCK(&stream->mutex); } _stream_destroy_workers(stream, &pool); device_switch_capturing(stream->dev, false); device_close(stream->dev); } void stream_loop_break(struct stream_t *stream) { atomic_store(&stream->proc->stop, true); } void stream_switch_slowdown(struct stream_t *stream, bool slowdown) { atomic_store(&stream->proc->slowdown, slowdown); } static bool _stream_wait_worker(struct stream_t *stream, struct workers_pool_t *pool, struct worker_t **ready_worker) { bool ok; A_MUTEX_LOCK(&pool->free_workers_mutex); A_COND_WAIT_TRUE(pool->free_workers, &pool->free_workers_cond, &pool->free_workers_mutex); A_MUTEX_UNLOCK(&pool->free_workers_mutex); *ready_worker = NULL; if (pool->oldest_worker && !atomic_load(&pool->oldest_worker->has_job) && pool->oldest_worker->buf_index >= 0) { *ready_worker = pool->oldest_worker; pool->oldest_worker = pool->oldest_worker->order_next; ok = true; // Воркер успел вовремя } else { for (unsigned number = 0; number < stream->dev->run->n_workers; ++number) { if ( !atomic_load(&pool->workers[number].has_job) && ( *ready_worker == NULL || (*ready_worker)->job_start_time < pool->workers[number].job_start_time ) ) { *ready_worker = &pool->workers[number]; break; } } assert(*ready_worker != NULL); ok = false; // Освободился воркер, получивший задание позже (или самый первый при самом первом захвате) } if ((*ready_worker)->job_failed) { *ready_worker = NULL; ok = false; } return ok; } static void _stream_expose_picture(struct stream_t *stream, unsigned buf_index) { # define PICTURE(_next) stream->dev->run->pictures[buf_index]._next A_MUTEX_LOCK(&stream->mutex); stream->picture.used = PICTURE(used); stream->picture.allocated = PICTURE(allocated); memcpy(stream->picture.data, PICTURE(data), stream->picture.used); stream->picture.grab_time = PICTURE(grab_time); stream->picture.encode_begin_time = PICTURE(encode_begin_time); stream->picture.encode_end_time = PICTURE(encode_end_time); stream->width = stream->dev->run->width; stream->height = stream->dev->run->height; atomic_store(&stream->updated, true); A_MUTEX_UNLOCK(&stream->mutex); # undef PICTURE } static void _stream_give_buf_to_worker(struct workers_pool_t *pool, struct worker_t *free_worker, unsigned buf_index) { if (pool->oldest_worker == NULL) { pool->oldest_worker = free_worker; pool->latest_worker = pool->oldest_worker; } else { if (free_worker->order_next) { free_worker->order_next->order_prev = free_worker->order_prev; } if (free_worker->order_prev) { free_worker->order_prev->order_next = free_worker->order_next; } free_worker->order_prev = pool->latest_worker; pool->latest_worker->order_next = free_worker; pool->latest_worker = free_worker; } pool->latest_worker->order_next = NULL; A_MUTEX_LOCK(&free_worker->has_job_mutex); free_worker->buf_index = buf_index; atomic_store(&free_worker->has_job, true); A_MUTEX_UNLOCK(&free_worker->has_job_mutex); A_COND_SIGNAL(&free_worker->has_job_cond); A_MUTEX_LOCK(&pool->free_workers_mutex); pool->free_workers -= 1; A_MUTEX_UNLOCK(&pool->free_workers_mutex); } static long double _stream_get_fluency_delay(struct device_t *dev, struct workers_pool_t *pool) { long double sum_comp_time = 0; long double avg_comp_time; long double min_delay; long double soft_delay; for (unsigned number = 0; number < dev->run->n_workers; ++number) { # define WORKER(_next) pool->workers[number]._next A_MUTEX_LOCK(&WORKER(last_comp_time_mutex)); if (WORKER(last_comp_time) > 0) { sum_comp_time += WORKER(last_comp_time); } A_MUTEX_UNLOCK(&WORKER(last_comp_time_mutex)); # undef WORKER } avg_comp_time = sum_comp_time / dev->run->n_workers; // Среднее время работы воркеров min_delay = avg_comp_time / dev->run->n_workers; // Среднее время работы размазывается на N воркеров if (dev->desired_fps > 0 && min_delay > 0) { // Искусственное время задержки на основе желаемого FPS, если включен --desired-fps soft_delay = ((long double) 1) / dev->desired_fps - sum_comp_time; return (min_delay > soft_delay ? min_delay : soft_delay); } return min_delay; } static int _stream_init_loop(struct stream_t *stream, struct workers_pool_t *pool) { int retval = -1; LOG_DEBUG("%s: stream->proc->stop = %d", __FUNCTION__, atomic_load(&stream->proc->stop)); while (!atomic_load(&stream->proc->stop)) { if ((retval = _stream_init(stream, pool)) < 0) { LOG_INFO("Sleeping %u seconds before new stream init ...", stream->dev->error_delay); sleep(stream->dev->error_delay); } else { break; } } return retval; } static int _stream_init(struct stream_t *stream, struct workers_pool_t *pool) { _stream_destroy_workers(stream, pool); device_switch_capturing(stream->dev, false); device_close(stream->dev); SEP_INFO('='); if (device_open(stream->dev) < 0) { goto error; } if (device_switch_capturing(stream->dev, true) < 0) { goto error; } encoder_prepare(pool->encoder, stream->dev); _stream_init_workers(stream, pool); return 0; error: device_close(stream->dev); return -1; } static void _stream_init_workers(struct stream_t *stream, struct workers_pool_t *pool) { LOG_INFO("Spawning %u workers ...", stream->dev->run->n_workers); atomic_store(&pool->workers_stop, false); A_CALLOC(pool->workers, stream->dev->run->n_workers); A_MUTEX_INIT(&pool->free_workers_mutex); A_COND_INIT(&pool->free_workers_cond); for (unsigned number = 0; number < stream->dev->run->n_workers; ++number) { # define WORKER(_next) pool->workers[number]._next pool->free_workers += 1; A_MUTEX_INIT(&WORKER(has_job_mutex)); atomic_init(&WORKER(has_job), false); A_COND_INIT(&WORKER(has_job_cond)); WORKER(number) = number; WORKER(proc_stop) = &stream->proc->stop; WORKER(workers_stop) = &pool->workers_stop; WORKER(free_workers_mutex) = &pool->free_workers_mutex; WORKER(free_workers) = &pool->free_workers; WORKER(free_workers_cond) = &pool->free_workers_cond; WORKER(dev) = stream->dev; WORKER(encoder) = pool->encoder; A_THREAD_CREATE(&WORKER(tid), _stream_worker_thread, (void *)&(pool->workers[number])); # undef WORKER } } static void *_stream_worker_thread(void *v_worker) { struct worker_t *worker = (struct worker_t *)v_worker; LOG_DEBUG("Hello! I am a worker #%u ^_^", worker->number); # ifdef WITH_WORKERS_GPIO_DEBUG # define WORKER_GPIO_DEBUG_BUSY digitalWrite(WORKERS_GPIO_DEBUG_START_PIN + worker->number, HIGH) # define WORKER_GPIO_DEBUG_FREE digitalWrite(WORKERS_GPIO_DEBUG_START_PIN + worker->number, LOW) pinMode(WORKERS_GPIO_DEBUG_START_PIN + worker->number, OUTPUT); WORKER_GPIO_DEBUG_FREE; # else # define WORKER_GPIO_DEBUG_BUSY # define WORKER_GPIO_DEBUG_FREE # endif while (!atomic_load(worker->proc_stop) && !atomic_load(worker->workers_stop)) { LOG_DEBUG("Worker %u waiting for a new job ...", worker->number); WORKER_GPIO_DEBUG_FREE; A_MUTEX_LOCK(&worker->has_job_mutex); A_COND_WAIT_TRUE(atomic_load(&worker->has_job), &worker->has_job_cond, &worker->has_job_mutex); A_MUTEX_UNLOCK(&worker->has_job_mutex); if (!atomic_load(worker->workers_stop)) { # define PICTURE(_next) worker->dev->run->pictures[worker->buf_index]._next LOG_DEBUG("Worker %u compressing JPEG from buffer %u ...", worker->number, worker->buf_index); WORKER_GPIO_DEBUG_BUSY; PICTURE(encode_begin_time) = get_now_monotonic(); if (encoder_compress_buffer(worker->encoder, worker->dev, worker->number, worker->buf_index) < 0) { worker->job_failed = false; } PICTURE(encode_end_time) = get_now_monotonic(); if (device_release_buffer(worker->dev, worker->buf_index) == 0) { worker->job_start_time = PICTURE(encode_begin_time); atomic_store(&worker->has_job, false); long double last_comp_time = PICTURE(encode_end_time) - worker->job_start_time; A_MUTEX_LOCK(&worker->last_comp_time_mutex); worker->last_comp_time = last_comp_time; A_MUTEX_UNLOCK(&worker->last_comp_time_mutex); LOG_VERBOSE("Compressed JPEG size=%zu; time=%0.3Lf; worker=%u; buffer=%u", PICTURE(used), last_comp_time, worker->number, worker->buf_index); } else { worker->job_failed = true; atomic_store(&worker->has_job, false); } # undef PICTURE } A_MUTEX_LOCK(worker->free_workers_mutex); *worker->free_workers += 1; A_MUTEX_UNLOCK(worker->free_workers_mutex); A_COND_SIGNAL(worker->free_workers_cond); } LOG_DEBUG("Bye-bye (worker %u)", worker->number); WORKER_GPIO_DEBUG_FREE; return NULL; } static void _stream_destroy_workers(struct stream_t *stream, struct workers_pool_t *pool) { if (pool->workers) { LOG_INFO("Destroying workers ..."); atomic_store(&pool->workers_stop, true); for (unsigned number = 0; number < stream->dev->run->n_workers; ++number) { # define WORKER(_next) pool->workers[number]._next A_MUTEX_LOCK(&WORKER(has_job_mutex)); atomic_store(&WORKER(has_job), true); // Final job: die A_MUTEX_UNLOCK(&WORKER(has_job_mutex)); A_COND_SIGNAL(&WORKER(has_job_cond)); A_THREAD_JOIN(WORKER(tid)); A_MUTEX_DESTROY(&WORKER(has_job_mutex)); A_COND_DESTROY(&WORKER(has_job_cond)); # undef WORKER } A_MUTEX_DESTROY(&pool->free_workers_mutex); A_COND_DESTROY(&pool->free_workers_cond); free(pool->workers); } pool->free_workers = 0; pool->workers = NULL; }