mirror of
https://github.com/MariaDB/server.git
synced 2025-01-16 12:02:42 +01:00
325 lines
9.1 KiB
C++
325 lines
9.1 KiB
C++
/* Copyright (c) 2008, 2015, Oracle and/or its affiliates. All rights reserved.
|
|
|
|
This program is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License, version 2.0,
|
|
as published by the Free Software Foundation.
|
|
|
|
This program is also distributed with certain software (including
|
|
but not limited to OpenSSL) that is licensed under separate terms,
|
|
as designated in a particular file or component or in included license
|
|
documentation. The authors of MySQL hereby grant you an additional
|
|
permission to link the program and your derivative works with the
|
|
separately licensed software that they have included with MySQL.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License, version 2.0, for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program; if not, write to the Free Software Foundation,
|
|
51 Franklin Street, Fifth Floor, Boston, MA 02110-1335 USA */
|
|
|
|
/**
|
|
@file storage/perfschema/pfs_events_waits.cc
|
|
Events waits data structures (implementation).
|
|
*/
|
|
|
|
#include "my_global.h"
|
|
#include "my_sys.h"
|
|
#include "pfs_global.h"
|
|
#include "pfs_instr_class.h"
|
|
#include "pfs_instr.h"
|
|
#include "pfs_user.h"
|
|
#include "pfs_host.h"
|
|
#include "pfs_account.h"
|
|
#include "pfs_events_waits.h"
|
|
#include "pfs_atomic.h"
|
|
#include "m_string.h"
|
|
|
|
ulong events_waits_history_long_size= 0;
|
|
/** Consumer flag for table EVENTS_WAITS_CURRENT. */
|
|
bool flag_events_waits_current= false;
|
|
/** Consumer flag for table EVENTS_WAITS_HISTORY. */
|
|
bool flag_events_waits_history= false;
|
|
/** Consumer flag for table EVENTS_WAITS_HISTORY_LONG. */
|
|
bool flag_events_waits_history_long= false;
|
|
/** Consumer flag for the global instrumentation. */
|
|
bool flag_global_instrumentation= false;
|
|
/** Consumer flag for the per thread instrumentation. */
|
|
bool flag_thread_instrumentation= false;
|
|
|
|
/** True if EVENTS_WAITS_HISTORY_LONG circular buffer is full. */
|
|
bool events_waits_history_long_full= false;
|
|
/** Index in EVENTS_WAITS_HISTORY_LONG circular buffer. */
|
|
volatile uint32 events_waits_history_long_index= 0;
|
|
/** EVENTS_WAITS_HISTORY_LONG circular buffer. */
|
|
PFS_events_waits *events_waits_history_long_array= NULL;
|
|
|
|
/**
|
|
Initialize table EVENTS_WAITS_HISTORY_LONG.
|
|
@param events_waits_history_long_sizing table sizing
|
|
*/
|
|
int init_events_waits_history_long(uint events_waits_history_long_sizing)
|
|
{
|
|
events_waits_history_long_size= events_waits_history_long_sizing;
|
|
events_waits_history_long_full= false;
|
|
PFS_atomic::store_u32(&events_waits_history_long_index, 0);
|
|
|
|
if (events_waits_history_long_size == 0)
|
|
return 0;
|
|
|
|
events_waits_history_long_array=
|
|
PFS_MALLOC_ARRAY(events_waits_history_long_size, sizeof(PFS_events_waits),
|
|
PFS_events_waits, MYF(MY_ZEROFILL));
|
|
|
|
return (events_waits_history_long_array ? 0 : 1);
|
|
}
|
|
|
|
/** Cleanup table EVENTS_WAITS_HISTORY_LONG. */
|
|
void cleanup_events_waits_history_long(void)
|
|
{
|
|
pfs_free(events_waits_history_long_array);
|
|
events_waits_history_long_array= NULL;
|
|
}
|
|
|
|
static inline void copy_events_waits(PFS_events_waits *dest,
|
|
const PFS_events_waits *source)
|
|
{
|
|
memcpy(dest, source, sizeof(PFS_events_waits));
|
|
}
|
|
|
|
/**
|
|
Insert a wait record in table EVENTS_WAITS_HISTORY.
|
|
@param thread thread that executed the wait
|
|
@param wait record to insert
|
|
*/
|
|
void insert_events_waits_history(PFS_thread *thread, PFS_events_waits *wait)
|
|
{
|
|
if (unlikely(events_waits_history_per_thread == 0))
|
|
return;
|
|
|
|
uint index= thread->m_waits_history_index;
|
|
|
|
/*
|
|
A concurrent thread executing TRUNCATE TABLE EVENTS_WAITS_CURRENT
|
|
could alter the data that this thread is inserting,
|
|
causing a potential race condition.
|
|
We are not testing for this and insert a possibly empty record,
|
|
to make this thread (the writer) faster.
|
|
This is ok, the readers of m_waits_history will filter this out.
|
|
*/
|
|
copy_events_waits(&thread->m_waits_history[index], wait);
|
|
|
|
index++;
|
|
if (index >= events_waits_history_per_thread)
|
|
{
|
|
index= 0;
|
|
thread->m_waits_history_full= true;
|
|
}
|
|
thread->m_waits_history_index= index;
|
|
}
|
|
|
|
/**
|
|
Insert a wait record in table EVENTS_WAITS_HISTORY_LONG.
|
|
@param wait record to insert
|
|
*/
|
|
void insert_events_waits_history_long(PFS_events_waits *wait)
|
|
{
|
|
if (unlikely(events_waits_history_long_size == 0))
|
|
return;
|
|
|
|
uint index= PFS_atomic::add_u32(&events_waits_history_long_index, 1);
|
|
|
|
index= index % events_waits_history_long_size;
|
|
if (index == 0)
|
|
events_waits_history_long_full= true;
|
|
|
|
/* See related comment in insert_events_waits_history. */
|
|
copy_events_waits(&events_waits_history_long_array[index], wait);
|
|
}
|
|
|
|
/** Reset table EVENTS_WAITS_CURRENT data. */
|
|
void reset_events_waits_current(void)
|
|
{
|
|
PFS_thread *pfs_thread= thread_array;
|
|
PFS_thread *pfs_thread_last= thread_array + thread_max;
|
|
|
|
for ( ; pfs_thread < pfs_thread_last; pfs_thread++)
|
|
{
|
|
PFS_events_waits *pfs_wait= pfs_thread->m_events_waits_stack;
|
|
PFS_events_waits *pfs_wait_last= pfs_wait + WAIT_STACK_SIZE;
|
|
|
|
for ( ; pfs_wait < pfs_wait_last; pfs_wait++)
|
|
pfs_wait->m_wait_class= NO_WAIT_CLASS;
|
|
}
|
|
}
|
|
|
|
/** Reset table EVENTS_WAITS_HISTORY data. */
|
|
void reset_events_waits_history(void)
|
|
{
|
|
PFS_thread *pfs_thread= thread_array;
|
|
PFS_thread *pfs_thread_last= thread_array + thread_max;
|
|
|
|
for ( ; pfs_thread < pfs_thread_last; pfs_thread++)
|
|
{
|
|
PFS_events_waits *wait= pfs_thread->m_waits_history;
|
|
PFS_events_waits *wait_last= wait + events_waits_history_per_thread;
|
|
|
|
pfs_thread->m_waits_history_index= 0;
|
|
pfs_thread->m_waits_history_full= false;
|
|
for ( ; wait < wait_last; wait++)
|
|
wait->m_wait_class= NO_WAIT_CLASS;
|
|
}
|
|
}
|
|
|
|
/** Reset table EVENTS_WAITS_HISTORY_LONG data. */
|
|
void reset_events_waits_history_long(void)
|
|
{
|
|
PFS_atomic::store_u32(&events_waits_history_long_index, 0);
|
|
events_waits_history_long_full= false;
|
|
|
|
PFS_events_waits *wait= events_waits_history_long_array;
|
|
PFS_events_waits *wait_last= wait + events_waits_history_long_size;
|
|
for ( ; wait < wait_last; wait++)
|
|
wait->m_wait_class= NO_WAIT_CLASS;
|
|
}
|
|
|
|
/** Reset table EVENTS_WAITS_SUMMARY_BY_THREAD_BY_EVENT_NAME data. */
|
|
void reset_events_waits_by_thread()
|
|
{
|
|
PFS_thread *thread= thread_array;
|
|
PFS_thread *thread_last= thread_array + thread_max;
|
|
PFS_account *account;
|
|
PFS_user *user;
|
|
PFS_host *host;
|
|
|
|
for ( ; thread < thread_last; thread++)
|
|
{
|
|
if (thread->m_lock.is_populated())
|
|
{
|
|
account= sanitize_account(thread->m_account);
|
|
user= sanitize_user(thread->m_user);
|
|
host= sanitize_host(thread->m_host);
|
|
aggregate_thread_waits(thread, account, user, host);
|
|
}
|
|
}
|
|
}
|
|
|
|
/** Reset table EVENTS_WAITS_SUMMARY_BY_ACCOUNT_BY_EVENT_NAME data. */
|
|
void reset_events_waits_by_account()
|
|
{
|
|
PFS_account *pfs= account_array;
|
|
PFS_account *pfs_last= account_array + account_max;
|
|
PFS_user *user;
|
|
PFS_host *host;
|
|
|
|
for ( ; pfs < pfs_last; pfs++)
|
|
{
|
|
if (pfs->m_lock.is_populated())
|
|
{
|
|
user= sanitize_user(pfs->m_user);
|
|
host= sanitize_host(pfs->m_host);
|
|
pfs->aggregate_waits(user, host);
|
|
}
|
|
}
|
|
}
|
|
|
|
/** Reset table EVENTS_WAITS_SUMMARY_BY_USER_BY_EVENT_NAME data. */
|
|
void reset_events_waits_by_user()
|
|
{
|
|
PFS_user *pfs= user_array;
|
|
PFS_user *pfs_last= user_array + user_max;
|
|
|
|
for ( ; pfs < pfs_last; pfs++)
|
|
{
|
|
if (pfs->m_lock.is_populated())
|
|
pfs->aggregate_waits();
|
|
}
|
|
}
|
|
|
|
/** Reset table EVENTS_WAITS_SUMMARY_BY_HOST_BY_EVENT_NAME data. */
|
|
void reset_events_waits_by_host()
|
|
{
|
|
PFS_host *pfs= host_array;
|
|
PFS_host *pfs_last= host_array + host_max;
|
|
|
|
for ( ; pfs < pfs_last; pfs++)
|
|
{
|
|
if (pfs->m_lock.is_populated())
|
|
pfs->aggregate_waits();
|
|
}
|
|
}
|
|
|
|
void reset_table_waits_by_table()
|
|
{
|
|
PFS_table_share *pfs= table_share_array;
|
|
PFS_table_share *pfs_last= pfs + table_share_max;
|
|
|
|
for ( ; pfs < pfs_last; pfs++)
|
|
{
|
|
if (pfs->m_lock.is_populated())
|
|
pfs->aggregate();
|
|
}
|
|
}
|
|
|
|
void reset_table_io_waits_by_table()
|
|
{
|
|
PFS_table_share *pfs= table_share_array;
|
|
PFS_table_share *pfs_last= pfs + table_share_max;
|
|
|
|
for ( ; pfs < pfs_last; pfs++)
|
|
{
|
|
if (pfs->m_lock.is_populated())
|
|
pfs->aggregate_io();
|
|
}
|
|
}
|
|
|
|
void reset_table_lock_waits_by_table()
|
|
{
|
|
PFS_table_share *pfs= table_share_array;
|
|
PFS_table_share *pfs_last= pfs + table_share_max;
|
|
|
|
for ( ; pfs < pfs_last; pfs++)
|
|
{
|
|
if (pfs->m_lock.is_populated())
|
|
pfs->aggregate_lock();
|
|
}
|
|
}
|
|
|
|
void reset_table_waits_by_table_handle()
|
|
{
|
|
PFS_table *pfs= table_array;
|
|
PFS_table *pfs_last= pfs + table_max;
|
|
|
|
for ( ; pfs < pfs_last; pfs++)
|
|
{
|
|
if (pfs->m_lock.is_populated())
|
|
pfs->sanitized_aggregate();
|
|
}
|
|
}
|
|
|
|
void reset_table_io_waits_by_table_handle()
|
|
{
|
|
PFS_table *pfs= table_array;
|
|
PFS_table *pfs_last= pfs + table_max;
|
|
|
|
for ( ; pfs < pfs_last; pfs++)
|
|
{
|
|
if (pfs->m_lock.is_populated())
|
|
pfs->sanitized_aggregate_io();
|
|
}
|
|
}
|
|
|
|
void reset_table_lock_waits_by_table_handle()
|
|
{
|
|
PFS_table *pfs= table_array;
|
|
PFS_table *pfs_last= pfs + table_max;
|
|
|
|
for ( ; pfs < pfs_last; pfs++)
|
|
{
|
|
if (pfs->m_lock.is_populated())
|
|
pfs->sanitized_aggregate_lock();
|
|
}
|
|
}
|
|
|