Tuesday, August 16, 2011

ORA-04031: unable to allocate 65560 bytes of shared memory ("large pool","unknown object","large pool","PX msg pool")

My Stats gathering job was failed and received the below alert.
Error details

ORA-12801: error signaled in parallel query server P055
ORA-12853: insufficient memory for PX buffers: current 66592K, max needed 178560K
ORA-04031: unable to allocate 65560 bytes of shared memory ("large pool","unknown object","large pool","PX msg pool")

I have checked database load & check the parallel session details. There is no parallel session running on database.

1.    This one is new job ?
Ans: No. It’s running fine for  last 2 years.
2.    Do you have faced the same issue before?
Ans: No.
3.    How longs it’s take to complete?
Ans: Less than 10 minutes
I asked to rerun the job & monitored the job session using Quest toad. That job spawned 97 parallel sessions and it’s failed again.
I have checked the PARALLEL_MAX_SERVERS & value of this parameter is 120.

That job procedure performs
1.    Drop the indexes
2.    Load the data
3.    Create the indexes
4.    Gather the statistics.
While gathering the statistics that job was failed and application team passed the arguments to procedure.
We have manually gathered the statistics successfully and ask application team to proceed the next step of the job.
But they don’t handling the exception for this job failure and the next job is depending upon the current job status.
we analyzed the procedure and it’s used below SQL code for stats gathering.
SQL Statement:
        || 'ownname => '         || '''' || p_towner || ''''
       || ',tabname => '        || '''' || p_tname  || ''''
       || ',method_opt => '   || '''FOR ALL COLUMNS SIZE AUTO'''
       || ',granularity => '   || '''ALL'''
       || ',cascade => TRUE'
       || ',degree => DBMS_STATS.DEFAULT_DEGREE ); END;';

ISSUE Details:
Above stats gather statement caused the issue on, “degree => DBMS_STATS.DEFAULT_DEGREE” value spawned multiple parallel threads (around 97 parallel sessions) on database. Due to this job was failed. That table having 2, 00,000 rows & why oracle optimizer spawned 97 parallel sessions for a small table. Sometimes Oracle 11g R1 optimizer behavior was very worst :-) 
Why it was previously running successfully? – Application team asked the root cause.

I have checked the oracle memory components value.  LARGE_POOL init parameter was not configured on the instance. While stats gathering oracle processes used the SHARED_POOL.

I have 2 solutions for this issue.

1)      It is recommended that application team need to look for the change in parameter  “degree => DBMS_STATS.DEFAULT_DEGREE” to “degree =>8” of their stats gathering procedure.
Before this change, I need to deploy these changes on non production and move to production with business approval & CM ticket.  So we couldn’t proceed these changes immediately.

2)     I will plan set the init parameter LARGE_POOL=300M on instance.
Unfortunately there is no memory on SGA components (MEMORY_TARGET and MEMORY_MAX_TARGET is fully allocated to other memory components). I don’t want to shrink the other memory components.

 I increased the MEMORY_TARGET, MEMORY_MAX_TARGET and LARGE_POOL value and recycled the database.

Now this job was running fine and this job don’t use any parallel session (its ran single thread).
Sometimes Oracle 11g R1 optimizer behavior was very worst :-)
Again there is a chance to happen the same issue. So we strongly recommended to developer.

“Need to control the parallelism on the proc or jobs when it is being configured. For small tables parallelism of 97 is too much and it can eat up most of the memory and CPU. So I recommended changing the “degree => DBMS_STATS.DEFAULT_DEGREE” to “degree =>8” of their stats gathering procedure”.

Hope this article helped to you. I am expecting your suggestions/feedback.

Best Regards
Rajabaskar Thangaraj


  1. awesome, thanks for sharing!!!

  2. two issues come to mind:
    -> why is dbms stats default degree hogging so many processes, even on a small table
    -> even when setting the large pool it may not get used

    That said, turning off parallelism does not solve the issue it just gets you around the current problem. Check

    Why is parallel_max_servers so high?
    Is it really using large pool? Did you try _px_use_large_pool=true?
    What is your setting for parallel_automatic_tuning and parallel_adaptive_multi_user?
    How does the setting of SGA_TARGET affect the use of large pool in your case? eg 0

  3. Klepjax, Thanks for writing in.. You were raising Good Questions..Need to digg more well and reply back asap. Thanks a lot