NonStop ODBC: how the connections (ODBC servers) are assigned to CPUs?
Posted
by Vladimir Dyuzhev
on Stack Overflow
See other posts from Stack Overflow
or by Vladimir Dyuzhev
Published on 2010-05-18T15:33:03Z
Indexed on
2010/05/22
4:00 UTC
Read the original article
Hit count: 372
We have an ODBC pool running on a NonStop server. The pool is connected to SQL/MX.
This pool is used by a few external Java applications, each of which has an JDBC pool connected to ODBC pool (e.g. 14 connections per application).
With time (after a few application recycles) we see an imbalance between CPUs -- some have 8 ODBC processes running, some only 5. That leads to CPU time imbalance too.
Up to this point we assumed that a CPU is assigned to ODBC process in round-robin fashion. That would maintain the number of ODBC processes more or less equally distributed. It's not the case though.
Is there any information on how ODBC pool decided which CPU to choose for every new allocated process? Does it look at CPU load? Available memory? Something else?
Sadly, even HP's own people (available to us, that is) couldn't answer those questions with certainty. :-(
© Stack Overflow or respective owner