NonStop ODBC: how the connections (ODBC servers) are assigned to CPUs?

Posted by Vladimir Dyuzhev on Stack Overflow See other posts from Stack Overflow or by Vladimir Dyuzhev
Published on 2010-05-18T15:33:03Z Indexed on 2010/05/22 4:00 UTC
Read the original article Hit count: 372

Filed under:
|
|

We have an ODBC pool running on a NonStop server. The pool is connected to SQL/MX.

This pool is used by a few external Java applications, each of which has an JDBC pool connected to ODBC pool (e.g. 14 connections per application).

With time (after a few application recycles) we see an imbalance between CPUs -- some have 8 ODBC processes running, some only 5. That leads to CPU time imbalance too.

Up to this point we assumed that a CPU is assigned to ODBC process in round-robin fashion. That would maintain the number of ODBC processes more or less equally distributed. It's not the case though.

Is there any information on how ODBC pool decided which CPU to choose for every new allocated process? Does it look at CPU load? Available memory? Something else?

Sadly, even HP's own people (available to us, that is) couldn't answer those questions with certainty. :-(

© Stack Overflow or respective owner

Related posts about odbc

Related posts about nonstop