- Global statistics collection is extremely expensive in terms of time and system resources
Published in Oracle Blog - By an user in Optimizer development team
This article will describe both of these issues and explain how you can address them both in Oracle Database 10gR2 and 11gR1.
Out of Range
Large tables are often decomposed into smaller pieces called partitions
in order to improve query performance and ease of data management. The
Oracle query optimizer relies on both the statistics of the entire table
(global statistics) and the statistics of the individual partitions
(partition statistics) to select a good execution plan for a SQL
statement. If the query needs to access only a single partition, the
optimizer uses only the statistics of the accessed partition. If the
query access more than one partition, it uses a combination of global
and partition statistics.
"Out of Range" means that the value supplied in a where clause
predicate is outside the domain of values represented by the [minimum,
maximum] column statistics. The optimizer prorates the selectivity based
on the distance between the predicate value and the maximum value
(assuming the value is higher than the max), that is, the farther the
value is from the maximum value, the lower the selectivity will be. This
situation occurs most frequently in tables that are range partitioned
by a date column, a new partition is added, and then queried while rows
are still being loaded in the new partition. The partition statistics
will be stale very quickly due to the continuous trickle feed load even
if the statistics get refreshed periodically. The maximum value known to
the optimizer is not correct leading to the "Out of Range" condition.
The under-estimation of selectivity often leads the query optimizer to
pick a sub optimal plan. For example, the query optimizer would pick an
index access path while a full scan is a better choice.
The "Out of Range" condition can be prevented by using the new copy
table statistics procedure available in Oracle Database10.2.0.4 and 11g.
This procedure copies the statistics of the source [sub] partition to
the destination [sub] partition. It also copies the statistics of the
dependent objects: columns, local (partitioned) indexes etc. It adjusts
the minimum and maximum values of the partitioning column as follows; it
uses the high bound partitioning value as the maximum value of the
first partitioning column (it is possible to have concatenated partition
columns) and high bound partitioning value of the previous partition as
the minimum value of the first partitioning column for range
partitioned table. It can optionally scale some of the other statistics
like the number of blocks, number of rows etc. of the destination
Assume we have a table called SALES that is ranged partitioned by quarter on the SALES_DATE column. At the end of every day data is loaded into latest partition. However, statistics are only gathered at the end of every quarter when the partition is fully loaded. Assuming global and partition level statistics (for all fully loaded partitions) are up to date, use the following steps in order to prevent getting a sub-optimal plan due to "out of range".1. Lock the table statistics using LOCK_TABLE_STATS procedure in DBMS_STATS. This is to avoid interference from auto statistics job.
2. Before beginning the initial load into each new partition (say SALES_Q4_2000) copy the statistics from the previous partition (say SALES_Q3_2000) using COPY_TABLE_STATS. You need to specify FORCE=>TRUE to override the statistics lock.
EXEC DBMS_STATS.COPY_TABLE_STATS ('SH', 'SALES', 'SALES_Q3_2000', 'SALES_Q4_2000', FORCE=>TRUE);
Expensive global statistics collection
data warehouse environment it is very common to do a bulk load directly
into one or more empty partitions. This will make the partition
statistics stale and may also make the global statistics stale.
Re-gathering statistics for the effected partitions and for the entire
table can be very time consuming. Traditionally, statistics collection
is done in a two-pass approach:
Oracle Database 11g, we avoid scanning the whole table when computing
global statistics by deriving the global statistics from the partition
statistics. Some of the statistics can be derived easily and accurately
from partition statistics. For example, number of rows at global level
is the sum of number of rows of partitions. Even global histogram can be
derived from partition histograms. But the number of distinct values
(NDV) of a column cannot be derived from partition level NDVs. So,
Oracle maintains another structure called a synopsis for each column at
the partition level. A synopsis can be considered as sample of distinct
values. The NDV can be accurately derived from synopses. We can also
merge multiple synopses into one. The global NDV is derived from the
synopsis generated by merging all of the partition level synopses. To
Incremental maintenance feature is disabled by default. It can be enabled by changing the INCREMENTAL table preference to true. It can also be enabled for a particular schema or at the database level. If you are interested in more details of the incremental maintenance feature, please refer to the following paper presented in SIGMOD 2008 and to our previous blog entry on new ndv gathering in 11g.Assume we have table called SALES that is range partitioned by day on the SALES_DATE column. At the end of every day data is loaded into latest partition and partition statistics are gathered. Global statistics are only gathered at the end of every month because gathering them is very time and resource intensive. Use the following steps in order to maintain global statistics after every load.
1 -Turn on incremental feature for the table.
Note: that the incremental maintenance feature was introduced in Oracle Database 11g Release 1. However, we also provide a solution in Oracle Database10g Release 2 (10.2.0.4) that simulates the same behavior. The 10g solution is a new value, 'APPROX_GLOBAL AND PARTITION' for the GRANULARITY parameter of the GATHER_TABLE_STATS procedures. It behaves the same as the incremental maintenance feature except that we don't update the NDV for non-partitioning columns and number of distinct keys of the index at the global level. For partitioned column we update the NDV as the sum of NDV at the partition levels. Also we set the NDV of columns of unique indexes as the number of rows of the table. In general, non-partitioning column NDV at the global level becomes stale less often. It may be possible to collect global statistics less frequently then the default (when table changes 10%) since approx_global option maintains most of the global statistics accurately.EXEC DBMS_STATS.GATHER_TABLE_STATS('SH','SALES');
Let's take a look at an example to see how you would effectively use the Oracle Database 10g approach.
After the data load is complete, gather statistics using DBMS_STATS.GATHER_TABLE_STATS for the last partition (say SALES_11FEB2009), specify granularity => 'APPROX_GLOBAL AND PARTITION'. It will collect statistics for the specified partition and derive global statistics from partition statistics (except for NDV as described before).
EXEC DBMS_STATS.GATHER_TABLE_STATS ('SH', 'SALES', 'SALES_11FEB2009', GRANULARITY => 'APPROX_GLOBAL AND PARTITION');It is necessary to install the one off patch for bug 8719831 if you are using the above features in 10.2.0.4 (patch 8877245) or in 126.96.36.199 (patch 8877251)