user10650209 wrote:I've seen this a few times when a cmt goes under extremely high load under solaris 10. Its possible the scheduler does not cope well with the high load. Being up to date on firmware / patches is a good start. Be aware if an old parallel garbage collection kicks in on a jvm this or worse can happen. prstat -mL might be an indicator if it happens again .... and if the SA decides to reboot tr to collect dump when he does so.
We're running Solaris 10 on a T3-2. The T3 is configured purely as a physical server; there are LDOM's created. We're running Oracle 11.2 databases on this server, and it's been running fine since May ( date of last reboot ). However, two days ago it 'hung'. There was no system panic, and from looking at Oracle AWR database reports, the load on the server went very high. At the same time as the load figure increased, the following messages started to appear in /var/adm/messages ...
Oct 23 05:59:21 pr2db1 ldc: [ID 522214 kern.warning] WARNING: ldc_close: (0xb) unregister failed, 11
Oct 23 06:02:22 pr2db1 last message repeated 157 times
Oct 23 06:02:23 pr2db1 ldc: [ID 522214 kern.warning] WARNING: ldc_close: (0xb) unregister failed, 11
Oct 23 06:09:03 pr2db1 last message repeated 350 times
The admin that was dealing with the issue resorted to rebooting the server, and it's been running okay for the last couple of days. I can't find too much about the above message. Has anyone got any pointers / thoughts ?