mirror of
https://github.com/Zygo/bees.git
synced 2025-06-17 01:56:16 +02:00
context: speed up orderly process termination
Quite often bees exceeds its service timeout for termination because it is waiting for a loop embedded in a Task to finish some long-running btrfs operation. This can cause bees to be aborted by SIGKILL before it can completely flush the hash table or save crawl state. There are only two important things SIGTERM does when bees terminates: 1. Save crawl progress 2. Flush out the hash table Everything else is automatically handled by the kernel when the process is terminated by SIGKILL, so we don't have to bother doing it ourselves. This can save considerable time at shutdown since we don't have to wait for every thread to reach a point where it becomes idle, or force loops to terminate by throwing exceptions, or check a condition every time we access a pointer. Instead, we need do only the things in the list above, and then call _exit() to clean up everything else. Hash table and crawl state writeback can happen in their background threads instead of the foreground one. Separate the "stop" method for these classes into "stop_request" and "stop_wait" so that these writebacks can run at the same time. Deprecate and remove all references to the BeesHalt exception, and remove several unnecessary checks for BeesContext::stop_requested. Pause the task queue instead of cancelling it, which preserves the crawl progress state and stops new Tasks from competing for iops and CPU during writeback. Signed-off-by: Zygo Blaxell <bees@furryterror.org>
This commit is contained in:
@ -155,22 +155,18 @@ BeesHashTable::flush_dirty_extents(bool slowly)
|
||||
if (flush_dirty_extent(extent_index)) {
|
||||
++wrote_extents;
|
||||
if (slowly) {
|
||||
if (m_stop_requested) {
|
||||
slowly = false;
|
||||
continue;
|
||||
}
|
||||
BEESNOTE("flush rate limited after extent #" << extent_index << " of " << m_extents << " extents");
|
||||
chrono::duration<double> sleep_time(m_flush_rate_limit.sleep_time(BLOCK_SIZE_HASHTAB_EXTENT));
|
||||
unique_lock<mutex> lock(m_stop_mutex);
|
||||
if (m_stop_requested) {
|
||||
BEESLOGDEBUG("Stop requested in hash table flush_dirty_extents");
|
||||
// This function is called by another thread with !slowly,
|
||||
// so we just get out of the way here.
|
||||
break;
|
||||
}
|
||||
m_stop_condvar.wait_for(lock, sleep_time);
|
||||
}
|
||||
}
|
||||
}
|
||||
if (!slowly) {
|
||||
BEESLOGINFO("Flushed " << wrote_extents << " of " << m_extents << " extents");
|
||||
}
|
||||
BEESLOGINFO("Flushed " << wrote_extents << " of " << m_extents << " hash table extents");
|
||||
return wrote_extents;
|
||||
}
|
||||
|
||||
@ -204,12 +200,27 @@ BeesHashTable::writeback_loop()
|
||||
m_dirty_condvar.wait(lock);
|
||||
}
|
||||
}
|
||||
|
||||
// The normal loop exits at the end of one iteration when stop requested,
|
||||
// but stop request will be in the middle of the loop, and some extents
|
||||
// will still be dirty. Run the flush loop again to get those.
|
||||
BEESNOTE("flushing hash table, round 2");
|
||||
BEESLOGDEBUG("Flushing hash table");
|
||||
flush_dirty_extents(false);
|
||||
|
||||
// If there were any Tasks still running, they may have updated
|
||||
// some hash table pages during the second flush. These updates
|
||||
// will be lost. The Tasks will be repeated on the next run because
|
||||
// they were not completed prior to the stop request, and the
|
||||
// Crawl progress was already flushed out before the Hash table
|
||||
// started writing, so nothing is really lost here.
|
||||
|
||||
catch_all([&]() {
|
||||
// trigger writeback on our way out
|
||||
#if 0
|
||||
// seems to trigger huge latency spikes
|
||||
BEESTOOLONG("unreadahead hash table size " << pretty(m_size));
|
||||
bees_unreadahead(m_fd, 0, m_size);
|
||||
BEESTOOLONG("unreadahead hash table size " <<
|
||||
pretty(m_size)); bees_unreadahead(m_fd, 0, m_size);
|
||||
#endif
|
||||
});
|
||||
BEESLOGDEBUG("Exited hash table writeback_loop");
|
||||
@ -794,7 +805,7 @@ BeesHashTable::~BeesHashTable()
|
||||
}
|
||||
|
||||
void
|
||||
BeesHashTable::stop()
|
||||
BeesHashTable::stop_request()
|
||||
{
|
||||
BEESNOTE("stopping BeesHashTable threads");
|
||||
BEESLOGDEBUG("Stopping BeesHashTable threads");
|
||||
@ -808,7 +819,11 @@ BeesHashTable::stop()
|
||||
unique_lock<mutex> dirty_lock(m_dirty_mutex);
|
||||
m_dirty_condvar.notify_all();
|
||||
dirty_lock.unlock();
|
||||
}
|
||||
|
||||
void
|
||||
BeesHashTable::stop_wait()
|
||||
{
|
||||
BEESNOTE("waiting for hash_prefetch thread");
|
||||
BEESLOGDEBUG("Waiting for hash_prefetch thread");
|
||||
m_prefetch_thread.join();
|
||||
@ -817,11 +832,5 @@ BeesHashTable::stop()
|
||||
BEESLOGDEBUG("Waiting for hash_writeback thread");
|
||||
m_writeback_thread.join();
|
||||
|
||||
if (m_cell_ptr && m_size) {
|
||||
BEESLOGDEBUG("Flushing hash table");
|
||||
BEESNOTE("flushing hash table");
|
||||
flush_dirty_extents(false);
|
||||
}
|
||||
|
||||
BEESLOGDEBUG("BeesHashTable stopped");
|
||||
}
|
||||
|
Reference in New Issue
Block a user