Search Results

Search found 78653 results on 3147 pages for 'performance object name s'.

Page 44/3147 | < Previous Page | 40 41 42 43 44 45 46 47 48 49 50 51  | Next Page >

  • Do functional generics exist and what is the correct name for them if they do?

    - by voroninp
    Consider the following generic class: public class EntityChangeInfo<EntityType,TEntityKey> { ChangeTypeEnum ChangeType {get;} TEntityKeyType EntityKey {get;} } Here EntityType unambiguously defines TEntityKeyType. So it would be nice to have some kind of types' map: public class EntityChangeInfo<EntityType,TEntityKey> with map < [ EntityType : Person -> TEntityKeyType : int] [ EntityType : Car -> TEntityKeyType : CarIdType ]> { ChangeTypeEnum ChangeType {get;} TEntityKeyType EntityKey {get;} } Another one example is: public class Foo<TIn> with map < [TIn : Person -> TOut1 : string, TOut2 : int, ..., TOutN : double ] [TIn : Car -> TOut1 : int, TOut2 :int, ..., TOutN : Price ] > { TOut1 Prop1 {get;set;} TOut2 Prop2 {get;set;} ... TOutN PropN {get;set;} } The reasonable question: how can this be interpreted by the compiler? Well, for me it is just the shortcut for two structurally similar classes: public sealed class Foo<Person> { string Prop1 {get;set;} int Prop2 {get;set;} ... double PropN {get;set;} } public sealed class Foo<Car> { int Prop1 {get;set;} int Prop2 {get;set;} ... Price PropN {get;set;} } But besides this we could imaging some update of the Foo<>: public class Foo<TIn> with map < [TIn : Person -> TOut1 : string, TOut2 : int, ..., TOutN : double ] [TIn : Car -> TOut1 : int, TOut2 :int, ..., TOutN : Price ] > { TOut1 Prop1 {get;set;} TOut2 Prop2 {get;set;} ... TOutN PropN {get;set;} public override string ToString() { return string.Format("prop1={0}, prop2={1},...propN={N-1}, Prop1, Prop2,...,PropN); } } This all can seem quite superficial but the idea came when I was designing the messages for our system. The very first class. Many messages with the same structure should be discriminated by the EntityType. So the question is whether such construct exists in any programming language?

    Read the article

  • Do functional generics exist or what is the correct name for them if they do?

    - by voroninp
    Consider the following generic class public class EntityChangeInfo<EntityType,TEntityKey> { ChangeTypeEnum ChangeType {get;} TEntityKeyType EntityKey {get;} } Here EntityType unambiguously defines TEntityKeyType. So it would be nice to have some kind of types' map public class EntityChangeInfo<EntityType,TEntityKey> with map < [ EntityType : Person -> TEntityKeyType : int] [ EntityType : Car -> TEntityKeyType : CarIdType ]> { ChangeTypeEnum ChangeType {get;} TEntityKeyType EntityKey {get;} } Another one example is: public class Foo<TIn> with map < [TIn : Person -> TOut1 : string, TOut2 : int, ..., TOutN : double ] [TIn : Car -> TOut1 : int, TOut2 :int, ..., TOutN : Price ] > { TOut1 Prop1 {get;set;} TOut2 Prop2 {get;set;} ... TOutN PropN {get;set;} } The reasonable question how this can be interpreted by the compiler? Well, for me it is just the sortcut for two structurally similar classes: public sealed class Foo<Person> { string Prop1 {get;set;} int Prop2 {get;set;} ... double PropN {get;set;} } public sealed class Foo<Car> { int Prop1 {get;set;} int Prop2 {get;set;} ... Price PropN {get;set;} } But besides this we could imaging some update of the Foo<: public class Foo<TIn> with map < [TIn : Person -> TOut1 : string, TOut2 : int, ..., TOutN : double ] [TIn : Car -> TOut1 : int, TOut2 :int, ..., TOutN : Price ] > { TOut1 Prop1 {get;set;} TOut2 Prop2 {get;set;} ... TOutN PropN {get;set;} public override string ToString() { return string.Format("prop1={0}, prop2={1},...propN={N-1}, Prop1, Prop2,...,PropN); } } This all can seem quite superficial but the idea came when I was designing the messages for our system. The very first class. Many messages with the same structrue should be discriminated by the EntityType. So the question is whether such construct exist in any programming language?

    Read the article

  • Testing smart card minidriver

    - by user352792
    when testing smart card minidriver in windows 7, got the following errors: "cmck exec Reconnect" always show that Testing through CAPI calls Submitting CSP PIN for reader \.\DMWZ ESAFE 0\ CryptAcquireContext - CRYPT_NEWKEYSET CryptGenKey Reconnecting CryptAcquireContext - CRYPT_DELETEKEYSET CryptAcquireContext failed unexpectedly d:\5429t\testsrc\dstest\security\core\credentials\smartcard\cmck\cmck\fnreconnect.cpp Line: 264 WIN32 0x80090016 Keyset does not exist. in windows xp, it always passed. i have no idea! this is my log. in XP: /* P:608 T:3380 8-30-203 CardAcquireContext(): BEGIN /* P:608 T:3380 8-30-203 CardAcquireContext(): SUCCESS /* P:608 T:3380 8-30-203 CardAcquireContext(): BEGIN /* P:608 T:3380 8-30-203 CardAcquireContext(): SUCCESS /* P:608 T:3380 8-31-750 CardAcquireContext(): BEGIN /* P:608 T:3380 8-31-765 CardAcquireContext(): SUCCESS /* P:608 T:3380 8-31-765 CardDeleteContext(): BEGIN /* P:608 T:3380 8-31-765 CardDeleteContext(): SUCCESS /* P:608 T:3380 8-31-765 CardAcquireContext(): BEGIN /* P:608 T:3380 8-31-765 CardAcquireContext(): SUCCESS /* P:608 T:3380 8-31-765 CardDeleteContext(): BEGIN /* P:608 T:3380 8-31-781 CardDeleteContext(): SUCCESS /* P:608 T:3380 8-31-781 CardAcquireContext(): BEGIN /* P:608 T:3380 8-31-781 CardAcquireContext(): SUCCESS /* P:608 T:3380 8-31-781 CardGetChallenge(): BEGIN /* P:608 T:3380 CardGetChallenge(): Challenge = CE568537C1BC9318 / / P:608 T:3380 8-31-781 CardGetChallenge(): SUCCESS /* P:608 T:3380 8-31-796 CardAuthenticateChallenge(): BEGIN /* P:608 T:3380 CardAuthenticateChallenge(): Response = B99E85F50E1F5C29 / / P:608 T:3380 8-31-796 CardAuthenticateChallenge(): SUCCESS /* P:608 T:3380 8-31-812 CardDeauthenticate(): BEGIN /* P:608 T:3380 8-31-812 CardDeauthenticate(): SUCCESS /* P:608 T:3380 8-31-812 CardAuthenticatePin(): BEGIN /* P:608 T:3380 CardAuthenticatePin(): User PIN = 0000 / / P:608 T:3380 8-31-828 CardAuthenticatePin(): SUCCESS /* P:608 T:3380 8-31-828 CardDeauthenticate(): BEGIN /* P:608 T:3380 8-31-843 CardDeauthenticate(): SUCCESS /* P:608 T:3380 8-31-843 CardDeleteContext(): BEGIN /* P:608 T:3380 8-31-843 CardDeleteContext(): SUCCESS /* P:608 T:3380 8-31-859 CardAcquireContext(): BEGIN /* P:608 T:3380 8-31-859 CardAcquireContext(): SUCCESS /* P:608 T:3380 8-31-859 CardAuthenticatePin(): BEGIN /* P:608 T:3380 CardAuthenticatePin(): User PIN = 0000 / / P:608 T:3380 8-31-875 CardAuthenticatePin(): SUCCESS /* P:608 T:3380 8-31-875 CardQueryCapabilities(): BEGIN /* P:608 T:3380 8-31-875 CardQueryCapabilities(): SUCCESS /* P:608 T:3380 8-31-890 CardAuthenticatePin(): BEGIN /* P:608 T:3380 CardAuthenticatePin(): User PIN = 0000 / / P:608 T:3380 8-31-906 CardAuthenticatePin(): SUCCESS /* P:608 T:3380 8-31-906 CardDeauthenticate(): BEGIN /* P:608 T:3380 8-31-921 CardDeauthenticate(): SUCCESS /* P:608 T:3380 8-31-921 CardDeleteContext(): BEGIN /* P:608 T:3380 8-31-921 CardDeleteContext(): SUCCESS /* P:608 T:3380 8-32-0 CardAcquireContext(): BEGIN /* P:608 T:3380 8-32-0 CardAcquireContext(): SUCCESS /* P:608 T:3380 8-32-0 CardReadFile(): BEGIN /* P:608 T:3380 CardReadFile(): Dir Name = ROOT, File Name = cardid / / P:608 T:3380 CardReadFile(): cardid = 34646533393531342D643465662D3432 / / P:608 T:3380 8-32-46 CardReadFile(): SUCCESS /* P:608 T:3380 8-32-62 CardReadFile(): BEGIN /* P:608 T:3380 CardReadFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardReadFile(): cardcf = 000000000000 / / P:608 T:3380 8-32-109 CardReadFile(): SUCCESS /* P:608 T:3380 8-32-109 CardReadFile(): BEGIN /* P:608 T:3380 CardReadFile(): Dir Name = mscp, File Name = cmapfile / / P:608 T:3380 8-32-187 CardReadFile(): BEGIN /* P:608 T:3380 CardReadFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardReadFile(): cardcf = 000000000000 / / P:608 T:3380 8-32-234 CardReadFile(): SUCCESS /* P:608 T:3380 8-32-250 CardAuthenticatePin(): BEGIN /* P:608 T:3380 CardAuthenticatePin(): User PIN = 0000 / / P:608 T:3380 8-32-265 CardAuthenticatePin(): SUCCESS /* P:608 T:3380 8-32-265 CardDeauthenticate(): BEGIN /* P:608 T:3380 8-32-281 CardDeauthenticate(): SUCCESS /* P:608 T:3380 8-32-281 CardReadFile(): BEGIN /* P:608 T:3380 CardReadFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardReadFile(): cardcf = 000000000000 / / P:608 T:3380 8-32-328 CardReadFile(): SUCCESS /* P:608 T:3380 8-32-343 CardQueryFreeSpace(): BEGIN /* P:608 T:3380 8-32-359 CardQueryFreeSpace(): SUCCESS /* P:608 T:3380 8-32-375 CardReadFile(): BEGIN /* P:608 T:3380 CardReadFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardReadFile(): cardcf = 000000000000 / / P:608 T:3380 8-32-421 CardReadFile(): SUCCESS /* P:608 T:3380 8-32-421 CardAuthenticatePin(): BEGIN /* P:608 T:3380 CardAuthenticatePin(): User PIN = 0000 / / P:608 T:3380 8-32-453 CardAuthenticatePin(): SUCCESS /* P:608 T:3380 8-32-453 CardWriteFile(): BEGIN /* P:608 T:3380 CardWriteFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardWriteFile(): cardcf = 000000000100 / / P:608 T:3380 8-32-531 CardWriteFile(): SUCCESS /* P:608 T:3380 8-32-531 CardWriteFile(): BEGIN /* P:608 T:3380 CardWriteFile(): Dir Name = mscp, File Name = cmapfile / / P:608 T:3380 CardWriteFile(): cmapfile = 660031006500300035003000300030002D0031003600380038002D0034006200380063002D0039006500300066002D003000310061006200300066006200340062003800660037000000000000000000010000000000 / / P:608 T:3380 8-32-921 CardWriteFile(): SUCCESS /* P:608 T:3380 8-32-921 CardWriteFile(): BEGIN /* P:608 T:3380 CardWriteFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardWriteFile(): cardcf = 000000000200 / / P:608 T:3380 8-33-0 CardWriteFile(): SUCCESS /* P:608 T:3380 8-33-0 CardWriteFile(): BEGIN /* P:608 T:3380 CardWriteFile(): Dir Name = mscp, File Name = cmapfile / / P:608 T:3380 CardWriteFile(): cmapfile = 660031006500300035003000300030002D0031003600380038002D0034006200380063002D0039006500300066002D003000310061006200300066006200340062003800660037000000000000000000030000000000 / / P:608 T:3380 8-33-109 CardWriteFile(): SUCCESS /* P:608 T:3380 8-33-125 CardQueryCapabilities(): BEGIN /* P:608 T:3380 8-33-125 CardQueryCapabilities(): SUCCESS /* P:608 T:3380 8-33-125 CardWriteFile(): BEGIN /* P:608 T:3380 CardWriteFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardWriteFile(): cardcf = 000001000200 / / P:608 T:3380 8-33-203 CardWriteFile(): SUCCESS /* P:608 T:3380 8-33-203 CardCreateContainer(): BEGIN /* P:608 T:3380 8-35-515 CardCreateContainer(): SUCCESS /* P:608 T:3380 8-35-531 CardWriteFile(): BEGIN /* P:608 T:3380 CardWriteFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardWriteFile(): cardcf = 000001000300 / / P:608 T:3380 8-35-609 CardWriteFile(): SUCCESS /* P:608 T:3380 8-35-609 CardWriteFile(): BEGIN /* P:608 T:3380 CardWriteFile(): Dir Name = mscp, File Name = cmapfile / / P:608 T:3380 CardWriteFile(): cmapfile = 660031006500300035003000300030002D0031003600380038002D0034006200380063002D0039006500300066002D003000310061006200300066006200340062003800660037000000000000000000030000040000 / / P:608 T:3380 8-35-734 CardWriteFile(): SUCCESS /* P:608 T:3380 8-35-734 CardGetContainerInfo(): BEGIN /* P:608 T:3380 8-35-796 CardGetContainerInfo(): SUCCESS /* P:608 T:5764 8-37-296 CardDeauthenticate(): BEGIN /* P:608 T:5764 8-37-312 CardDeauthenticate(): SUCCESS /* P:608 T:3380 8-37-312 CardReadFile(): BEGIN /* P:608 T:3380 CardReadFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardReadFile(): cardcf = 000001000300 / / P:608 T:3380 8-37-375 CardReadFile(): SUCCESS /* P:608 T:3380 8-37-375 CardReadFile(): BEGIN /* P:608 T:3380 CardReadFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardReadFile(): cardcf = 000001000300 / / P:608 T:3380 8-37-437 CardReadFile(): SUCCESS /* P:608 T:3380 8-37-437 CardAuthenticatePin(): BEGIN /* P:608 T:3380 CardAuthenticatePin(): User PIN = 0000 / / P:608 T:3380 8-37-468 CardAuthenticatePin(): SUCCESS /* P:608 T:3380 8-37-484 CardWriteFile(): BEGIN /* P:608 T:3380 CardWriteFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardWriteFile(): cardcf = 000001000400 / / P:608 T:3380 8-37-546 CardWriteFile(): SUCCESS /* P:608 T:3380 8-37-562 CardDeleteFile(): BEGIN /* P:608 T:3380 CardDeleteFile(): Dir Name = mscp, File Name = ksc00 / / P:608 T:3380 8-37-625 CardDeleteFile(): SCARD_E_FILE_NOT_FOUND (0x80100024) /* P:608 T:3380 CardDeleteFile(): FAILED /* P:608 T:3380 8-37-625 CardReadFile(): BEGIN /* P:608 T:3380 CardReadFile(): Dir Name = mscp, File Name = cmapfile / / P:608 T:3380 CardReadFile(): cmapfile = 660031006500300035003000300030002D0031003600380038002D0034006200380063002D0039006500300066002D003000310061006200300066006200340062003800660037000000000000000000030000040000 / / P:608 T:3380 8-37-718 CardReadFile(): SUCCESS /* P:608 T:3380 8-37-718 CardWriteFile(): BEGIN /* P:608 T:3380 CardWriteFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardWriteFile(): cardcf = 000001000500 / / P:608 T:3380 8-37-796 CardWriteFile(): SUCCESS /* P:608 T:3380 8-37-796 CardDeleteFile(): BEGIN /* P:608 T:3380 CardDeleteFile(): Dir Name = mscp, File Name = kxc00 / / P:608 T:3380 8-37-875 CardDeleteFile(): SCARD_E_FILE_NOT_FOUND (0x80100024) /* P:608 T:3380 CardDeleteFile(): FAILED /* P:608 T:3380 8-37-875 CardWriteFile(): BEGIN /* P:608 T:3380 CardWriteFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardWriteFile(): cardcf = 000002000500 / / P:608 T:3380 8-37-953 CardWriteFile(): SUCCESS /* P:608 T:3380 8-37-953 CardDeleteContainer(): BEGIN /* P:608 T:3380 8-38-578 CardDeleteContainer(): SUCCESS /* P:608 T:3380 8-38-593 CardReadFile(): BEGIN /* P:608 T:3380 CardReadFile(): Dir Name = mscp, File Name = cmapfile / / P:608 T:3380 CardReadFile(): cmapfile = 660031006500300035003000300030002D0031003600380038002D0034006200380063002D0039006500300066002D003000310061006200300066006200340062003800660037000000000000000000030000040000 / / P:608 T:3380 8-38-687 CardReadFile(): SUCCESS /* P:608 T:3380 8-38-687 CardWriteFile(): BEGIN /* P:608 T:3380 CardWriteFile(): Dir Name = ROOT, File Name = cardcf / / P:608 T:3380 CardWriteFile(): cardcf = 000002000600 / / P:608 T:3380 8-38-781 CardWriteFile(): SUCCESS /* P:608 T:3380 8-38-781 CardWriteFile(): BEGIN /* P:608 T:3380 CardWriteFile(): Dir Name = mscp, File Name = cmapfile / / P:608 T:3380 CardWriteFile(): cmapfile = 0000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000 / / P:608 T:3380 8-38-906 CardWriteFile(): SUCCESS /* P:608 T:5764 8-40-406 CardDeauthenticate(): BEGIN /* P:608 T:5764 8-40-421 CardDeauthenticate(): SUCCESS /* P:608 T:3380 8-40-671 CardDeleteContext(): BEGIN /* P:608 T:3380 8-40-687 CardDeleteContext(): SUCCESS in windows 7: /* P:3368 T:3800 17-39-515 CardAcquireContext(): BEGIN /* P:3368 T:3800 17-39-515 CardAcquireContext(): SUCCESS /* P:3368 T:3800 17-39-515 CardAcquireContext(): BEGIN /* P:3368 T:3800 17-39-515 CardAcquireContext(): SUCCESS /* P:3368 T:3800 17-39-531 CardAcquireContext(): BEGIN /* P:3368 T:3800 17-39-531 CardAcquireContext(): SUCCESS /* P:3368 T:3800 17-39-531 CardAcquireContext(): BEGIN /* P:3368 T:3800 17-39-531 CardAcquireContext(): SUCCESS /* P:3368 T:3800 17-41-187 CardAcquireContext(): BEGIN /* P:3368 T:3800 17-41-187 CardAcquireContext(): SUCCESS /* P:3368 T:3800 17-41-187 CardDeleteContext(): BEGIN /* P:3368 T:3800 17-41-187 CardDeleteContext(): SUCCESS /* P:3368 T:3800 17-41-187 CardAcquireContext(): BEGIN /* P:3368 T:3800 17-41-187 CardAcquireContext(): SUCCESS /* P:3368 T:3800 17-41-187 CardDeleteContext(): BEGIN /* P:3368 T:3800 17-41-203 CardDeleteContext(): SUCCESS /* P:3368 T:3800 17-41-203 CardAcquireContext(): BEGIN /* P:3368 T:3800 17-41-203 CardAcquireContext(): SUCCESS /* P:3368 T:3800 17-41-203 CardDeleteContext(): BEGIN /* P:3368 T:3800 17-41-203 CardDeleteContext(): SUCCESS /* P:3368 T:3800 17-41-203 CardAcquireContext(): BEGIN /* P:3368 T:3800 17-41-203 CardAcquireContext(): SUCCESS /* P:3368 T:3800 17-41-218 CardDeleteContext(): BEGIN /* P:3368 T:3800 17-41-218 CardDeleteContext(): SUCCESS /* P:3368 T:3800 17-41-218 CardAcquireContext(): BEGIN /* P:3368 T:3800 17-41-218 CardAcquireContext(): SUCCESS /* P:3368 T:3800 17-41-218 CardGetChallenge(): BEGIN /* P:3368 T:3800 CardGetChallenge(): Challenge = BF830855CDCA4F0D / / P:3368 T:3800 17-41-234 CardGetChallenge(): SUCCESS /* P:3368 T:3800 17-41-234 CardAuthenticateChallenge(): BEGIN /* P:3368 T:3800 CardAuthenticateChallenge(): Response = A2DB6F882D402D94 / / P:3368 T:3800 17-41-234 CardAuthenticateChallenge(): SUCCESS /* P:3368 T:3800 17-41-234 CardDeauthenticate(): BEGIN /* P:3368 T:3800 17-41-250 CardDeauthenticate(): SUCCESS /* P:3368 T:3800 17-41-250 CardAuthenticatePin(): BEGIN /* P:3368 T:3800 CardAuthenticatePin(): User PIN = 0000 / / P:3368 T:3800 17-41-265 CardAuthenticatePin(): SUCCESS /* P:3368 T:3800 17-41-265 CardDeauthenticate(): BEGIN /* P:3368 T:3800 17-41-265 CardDeauthenticate(): SUCCESS /* P:3368 T:3800 17-41-265 CardDeleteContext(): BEGIN /* P:3368 T:3800 17-41-281 CardDeleteContext(): SUCCESS /* P:3368 T:3800 17-41-281 CardAcquireContext(): BEGIN /* P:3368 T:3800 17-41-281 CardAcquireContext(): SUCCESS /* P:3368 T:3800 17-41-281 CardAuthenticatePin(): BEGIN /* P:3368 T:3800 CardAuthenticatePin(): User PIN = 0000 / / P:3368 T:3800 17-41-296 CardAuthenticatePin(): SUCCESS /* P:3368 T:3800 17-41-296 CardQueryCapabilities(): BEGIN /* P:3368 T:3800 17-41-296 CardQueryCapabilities(): SUCCESS /* P:3368 T:3800 17-41-296 CardAuthenticatePin(): BEGIN /* P:3368 T:3800 CardAuthenticatePin(): User PIN = 0000 / / P:3368 T:3800 17-41-312 CardAuthenticatePin(): SUCCESS /* P:3368 T:3800 17-41-312 CardDeauthenticate(): BEGIN /* P:3368 T:3800 17-41-328 CardDeauthenticate(): SUCCESS /* P:3368 T:3800 17-41-328 CardDeleteContext(): BEGIN /* P:3368 T:3800 17-41-328 CardDeleteContext(): SUCCESS /* P:3368 T:3800 17-41-359 CardAcquireContext(): BEGIN /* P:3368 T:3800 17-41-359 CardAcquireContext(): SUCCESS /* P:3368 T:3800 17-41-359 CardReadFile(): BEGIN /* P:3368 T:3800 CardReadFile(): Dir Name = ROOT, File Name = cardid / / P:3368 T:3800 CardReadFile(): cardid = 34363438653733652D346430342D3463 / / P:3368 T:3800 17-41-406 CardReadFile(): SUCCESS /* P:3368 T:3800 17-41-406 CardReadFile(): BEGIN /* P:3368 T:3800 CardReadFile(): Dir Name = ROOT, File Name = cardcf / / P:3368 T:3800 CardReadFile(): cardcf = 000000000000 / / P:3368 T:3800 17-41-453 CardReadFile(): SUCCESS /* P:3368 T:3800 17-41-453 CardReadFile(): BEGIN /* P:3368 T:3800 CardReadFile(): Dir Name = mscp, File Name = cmapfile / / P:3368 T:3800 17-41-531 CardReadFile(): BEGIN /* P:3368 T:3800 CardReadFile(): Dir Name = ROOT, File Name = cardcf / / P:3368 T:3800 CardReadFile(): cardcf = 000000000000 / / P:3368 T:3800 17-41-593 CardReadFile(): SUCCESS /* P:3368 T:3800 17-41-593 CardAuthenticatePin(): BEGIN /* P:3368 T:3800 CardAuthenticatePin(): User PIN = 0000 / / P:3368 T:3800 17-41-609 CardAuthenticatePin(): SUCCESS /* P:3368 T:3800 17-41-609 CardDeauthenticate(): BEGIN /* P:3368 T:3800 17-41-609 CardDeauthenticate(): SUCCESS /* P:3368 T:3800 17-41-609 CardDeleteContext(): BEGIN /* P:3368 T:3800 17-41-625 CardDeleteContext(): SUCCESS /* P:3368 T:3800 17-41-625 CardAcquireContext(): BEGIN /* P:3368 T:3800 17-41-625 CardAcquireContext(): SUCCESS /* P:3368 T:3800 17-41-625 CardReadFile(): BEGIN /* P:3368 T:3800 CardReadFile(): Dir Name = ROOT, File Name = cardid / / P:3368 T:3800 CardReadFile(): cardid = 34363438653733652D346430342D3463 / / P:3368 T:3800 17-41-671 CardReadFile(): SUCCESS /* P:3368 T:3800 17-41-687 CardReadFile(): BEGIN /* P:3368 T:3800 CardReadFile(): Dir Name = ROOT, File Name = cardcf / / P:3368 T:3800 CardReadFile(): cardcf = 000000000000 / / P:3368 T:3800 17-41-734 CardReadFile(): SUCCESS /* P:3368 T:3800 17-41-734 CardQueryFreeSpace(): BEGIN /* P:3368 T:3800 17-41-750 CardQueryFreeSpace(): SUCCESS /* P:3368 T:3800 17-41-750 CardAuthenticatePin(): BEGIN /* P:3368 T:3800 CardAuthenticatePin(): User PIN = 0000 / / P:3368 T:3800 17-41-765 CardAuthenticatePin(): SUCCESS /* P:3368 T:3800 17-41-765 CardWriteFile(): BEGIN /* P:3368 T:3800 CardWriteFile(): Dir Name = ROOT, File Name = cardcf / / P:3368 T:3800 CardWriteFile(): cardcf = 000000000100 / / P:3368 T:3800 17-41-828 CardWriteFile(): SUCCESS /* P:3368 T:3800 17-41-828 CardWriteFile(): BEGIN /* P:3368 T:3800 CardWriteFile(): Dir Name = mscp, File Name = cmapfile / / P:3368 T:3800 CardWriteFile(): cmapfile = 370062003800640030006200390031002D0063003600650064002D0034003000650033002D0062006100610037002D006200620032003800640063003800610035003300330032000000000000000000010000000000 / / P:3368 T:3800 17-42-218 CardWriteFile(): SUCCESS /* P:3368 T:3800 17-42-234 CardWriteFile(): BEGIN /* P:3368 T:3800 CardWriteFile(): Dir Name = ROOT, File Name = cardcf / / P:3368 T:3800 CardWriteFile(): cardcf = 000000000200 / / P:3368 T:3800 17-42-296 CardWriteFile(): SUCCESS /* P:3368 T:3800 17-42-296 CardWriteFile(): BEGIN /* P:3368 T:3800 CardWriteFile(): Dir Name = mscp, File Name = cmapfile / / P:3368 T:3800 CardWriteFile(): cmapfile = 370062003800640030006200390031002D0063003600650064002D0034003000650033002D0062006100610037002D006200620032003800640063003800610035003300330032000000000000000000030000000000 / / P:3368 T:3800 17-42-390 CardWriteFile(): SUCCESS /* P:3368 T:3800 17-42-406 CardQueryCapabilities(): BEGIN /* P:3368 T:3800 17-42-406 CardQueryCapabilities(): SUCCESS /* P:3368 T:3800 17-42-406 CardWriteFile(): BEGIN /* P:3368 T:3800 CardWriteFile(): Dir Name = ROOT, File Name = cardcf / / P:3368 T:3800 CardWriteFile(): cardcf = 000001000200 / / P:3368 T:3800 17-42-468 CardWriteFile(): SUCCESS /* P:3368 T:3800 17-42-468 CardCreateContainer(): BEGIN /* P:3368 T:3800 17-48-421 CardCreateContainer(): SUCCESS /* P:3368 T:3800 17-48-437 CardWriteFile(): BEGIN /* P:3368 T:3800 CardWriteFile(): Dir Name = ROOT, File Name = cardcf / / P:3368 T:3800 CardWriteFile(): cardcf = 000001000300 / / P:3368 T:3800 17-48-484 CardWriteFile(): SUCCESS /* P:3368 T:3800 17-48-500 CardWriteFile(): BEGIN /* P:3368 T:3800 CardWriteFile(): Dir Name = mscp, File Name = cmapfile / / P:3368 T:3800 CardWriteFile(): cmapfile = 370062003800640030006200390031002D0063003600650064002D0034003000650033002D0062006100610037002D006200620032003800640063003800610035003300330032000000000000000000030000040000 / / P:3368 T:3800 17-48-593 CardWriteFile(): SUCCESS /* P:3368 T:3800 17-48-593 CardGetContainerInfo(): BEGIN /* P:3368 T:3800 17-48-640 CardGetContainerInfo(): SUCCESS /* P:3368 T:288 17-50-140 CardDeauthenticate(): BEGIN /* P:3368 T:288 17-50-140 CardDeauthenticate(): SUCCESS /* P:3368 T:3800 17-50-140 CardReadFile(): BEGIN /* P:3368 T:3800 CardReadFile(): Dir Name = ROOT, File Name = cardid / / P:3368 T:3800 CardReadFile(): cardid = 34363438653733652D346430342D3463 / / P:3368 T:3800 17-50-187 CardReadFile(): SUCCESS /* P:3368 T:3800 17-50-187 CardReadFile(): BEGIN /* P:3368 T:3800 CardReadFile(): Dir Name = ROOT, File Name = cardcf / / P:3368 T:3800 CardReadFile(): cardcf = 000001000300 / / P:3368 T:3800 17-50-234 CardReadFile(): SUCCESS /* P:3368 T:3800 17-50-234 CardReadFile(): BEGIN /* P:3368 T:3800 CardReadFile(): Dir Name = ROOT, File Name = cardid / / P:3368 T:3800 CardReadFile(): cardid = 34363438653733652D346430342D3463 / / P:3368 T:3800 17-50-296 CardReadFile(): SUCCESS /* P:3368 T:3800 17-50-296 CardReadFile(): BEGIN /* P:3368 T:3800 CardReadFile(): Dir Name = ROOT, File Name = cardid / / P:3368 T:3800 CardReadFile(): cardid = 34363438653733652D346430342D3463 / / P:3368 T:3800 17-50-343 CardReadFile(): SUCCESS Comparing the two logs, it seems that in win 7 cmck always read file, read file, read file... and fail, never get into CardDeleteContainer or CardWriteFile :( Please help me!!!! Many thanks!

    Read the article

  • Ruby/Rails - Add records to an object with each loop iteration / Object vs Arrays

    - by ChrisWesAllen
    I'm trying to figure out how to add records to an existing object for each iteration of a loop. I'm having a hard time discovering the difference between an object and an array. I have this @events = Event.find(1) @loops = Choices.find(:all, :limit => 5) #so loop for 5 instances of choice model for loop in @loops @events = Event.find(:all,:conditions => ["event.id = ?", loop.event_id ]) end I'm trying to add a new events to the existing @events object based on the id of whatever the loop variable is. But the ( = ) operator just creates a new instance of the @events object. I tried ( += ) and ( << ) as operators but got the error "You have a nil object when you didn't expect it! You might have expected an instance of Array. The error occurred while evaluating nil" I tried created an array events = [] events << Event.find(1) @loops = Choices.find(:all, :limit => 5) #so loop for 5 instances of choice model for loop in @loops events << Event.find(:all,:conditions => ["event.id = ?", loop.event_id ]) end But I dont know how to call that arrays attributes within the view With objects I was able do create a loop within the view and call all the attributes of that object as well... <table> <% for event in @events %> <tr> <td><%= link_to event.title, event %></td> <td><%= event.start_date %></td> <td><%= event.price %></td> </tr> <% end %> </table> How could i do this with an array set? So the questions are 1) Whats the difference between arrays and objects? 2) Is there a way to add into the existing object for each iteration? 3) If I use an array, is there a way to call the attributes for each array record within the view?

    Read the article

  • Application Performance Episode 2: Announcing the Judges!

    - by Michaela Murray
    The story so far… We’re writing a new book for ASP.NET developers, and we want you to be a part of it! If you work with ASP.NET applications, and have top tips, hard-won lessons, or sage advice for avoiding, finding, and fixing performance problems, we want to hear from you! And if your app uses SQL Server, even better – interaction with the database is critical to application performance, so we’re looking for database top tips too. There’s a Microsoft Surface apiece for the person who comes up with the best tip for SQL Server and the best tip for .NET. Of course, if your suggestion is selected for the book, you’ll get full credit, by name, Twitter handle, GitHub repository, or whatever you like. To get involved, just email your nuggets of performance wisdom to [email protected] – there are examples of what we’re looking for and full competition details at Application Performance: The Best of the Web. Enter the judges… As mentioned in my last blogpost, we have a mystery panel of celebrity judges lined up to select the prize-winning performance pointers. We’re now ready to reveal their secret identities! Judging your ASP.NET  tips will be: Jean-Phillippe Gouigoux, MCTS/MCPD Enterprise Architect and MVP Connected System Developer. He’s a board member at French software company MGDIS, and teaches algorithms, security, software tests, and ALM at the Université de Bretagne Sud. Jean-Philippe also lectures at IT conferences and writes articles for programming magazines. His book Practical Performance Profiling is published by Simple-Talk. Nik Molnar,  a New Yorker, ASP Insider, and co-founder of Glimpse, an open source ASP.NET diagnostics and debugging tool. Originally from Florida, Nik specializes in web development, building scalable, client-centric solutions. In his spare time, Nik can be found cooking up a storm in the kitchen, hanging with his wife, speaking at conferences, and working on other open source projects. Mitchel Sellers, Microsoft C# and DotNetNuke MVP. Mitchel is an experienced software architect, business leader, public speaker, and educator. He works with companies across the globe, as CEO of IowaComputerGurus Inc. Mitchel writes technical articles for online and print publications and is the author of Professional DotNetNuke Module Programming. He frequently answers questions on StackOverflow and MSDN and is an active participant in the .NET and DotNetNuke communities. Clive Tong, Software Engineer at Red Gate. In previous roles, Clive spent a lot of time working with Common LISP and enthusing about functional languages, and he’s worked with managed languages since before his first real job (which was a long time ago). Long convinced of the productivity benefits of managed languages, Clive is very interested in getting good runtime performance to keep managed languages practical for real-world development. And our trio of SQL Server specialists, ready to select your top suggestion, are (drumroll): Rodney Landrum, a SQL Server MVP who writes regularly about Integration Services, Analysis Services, and Reporting Services. He’s authored SQL Server Tacklebox, three Reporting Services books, and contributes regularly to SQLServerCentral, SQL Server Magazine, and Simple–Talk. His day job involves overseeing a large SQL Server infrastructure in Orlando. Grant Fritchey, Product Evangelist at Red Gate and SQL Server MVP. In an IT career spanning more than 20 years, Grant has written VB, VB.NET, C#, and Java. He’s been working with SQL Server since version 6.0. Grant volunteers with the Editorial Committee at PASS and has written books for Apress and Simple-Talk. Jonathan Allen, leader and founder of the PASS SQL South West user group. He’s been working with SQL Server since 1999 and enjoys performance tuning, development, and using SQL Server for business solutions. He’s spoken at SQLBits and SQL in the City, as well as local user groups across the UK. He’s also a moderator at ask.sqlservercentral.com.

    Read the article

  • Performance penalty of typecasting and boxing/unboxing types in C# when storing generic values

    - by kitsune
    I have a set-up similar to WPF's DependencyProperty and DependencyObject system. My properties however are generic. A BucketProperty has a static GlobalIndex (defined in BucketPropertyBase) which tracks all BucketProperties. A Bucket can have many BucketProperties of any type. A Bucket saves and gets the actual values of these BucketProperties... now my question is, how to deal with the storage of these values, and what is the penalty of using a typecasting when retrieving them? I currently use an array of BucketEntries that save the property values as simple objects. Is there any better way of saving and returning these values? Beneath is a simpliefied version: public class BucketProperty<T> : BucketPropertyBase { } public class Bucket { private BucketEntry[] _bucketEntries; public void SaveValue<T>(BucketProperty<T> property, T value) { SaveBucketEntry(property.GlobalIndex, value) } public T GetValue<T>(BucketProperty<T> property) { return (T)FindBucketEntry(property.GlobalIndex).Value; } } public class BucketEntry { private object _value; private uint _index; public BucketEntry(uint globalIndex, object value) { ... } }

    Read the article

  • Silverlight combobox performance issue

    - by Vinzz
    Hi, I'm facing a performance issue with a crowded combobox (5000 items). Rendering of the drop down list is really slow (as if it was computing all items before showing any). Do you have any trick to make this dropdown display lazy? Xaml code: <Grid x:Name="LayoutRoot"> <StackPanel Orientation="Horizontal" Width="200" Height="20"> <TextBlock>Test Combo </TextBlock> <ComboBox x:Name="fooCombo" Margin="5,0,0,0"></ComboBox> </StackPanel> </Grid> code behind: public MainPage() { InitializeComponent(); List<string> li = new List<string>(); int Max = 5000; for (int i = 0; i < Max; ++i) li.Add("Item - " + i); fooCombo.ItemsSource = li; }

    Read the article

  • SQLite self-join performance

    - by Derk
    What I essentially want, is to retreive all features and values of products which have a particular feature and value. For example: I want to know all available hard drive sizes of products that have an Intel processor. I have three tables: product_to_value (product_id, feature_id, value_id) features (id, value) // for example Processor family, Storage size, etc. values (id, value) // for example Intel, 60GB, etc The simplified query I have now: SELECT features.name, featurevalues.name, featurevalues.value FROM products, products as prod2, features, features as feat2, values, values as val2 WHERE products.feature = features.id AND products.value = values.id AND products.product = prod2.product AND prod2.feature_id = feat2.id AND prod2.value_id = val2.id AND features.id = ? AND feat2.id = ? All columns have an index. I am using SQLite. The problem is that it's very slow (70ms per query, without the self-join it's <1ms). Is there a smarter way to fetch data like this? Or is this too much to ask from SQLite? I personally think I am simply overlooking something, as I am quite new to SQLite.

    Read the article

  • MySQL MyISAM table performance... painfully, painfully slow

    - by Salman A
    I've got a table structure that can be summarized as follows: pagegroup * pagegroupid * name has 3600 rows page * pageid * pagegroupid * data references pagegroup; has 10000 rows; can have anything between 1-700 rows per pagegroup; the data column is of type mediumtext and the column contains 100k - 200kbytes data per row userdata * userdataid * pageid * column1 * column2 * column9 references page; has about 300,000 rows; can have about 1-50 rows per page The above structure is pretty straight forwad, the problem is that that a join from userdata to page group is terribly, terribly slow even though I have indexed all columns that should be indexed. The time needed to run a query for such a join (userdata inner_join page inner_join pagegroup) exceeds 3 minutes. This is terribly slow considering the fact that I am not selecting the data column at all. Example of the query that takes too long: SELECT userdata.column1, pagegroup.name FROM userdata INNER JOIN page USING( pageid ) INNER JOIN pagegroup USING( pagegroupid ) Please help by explaining why does it take so long and what can i do to make it faster. Edit #1 Explain returns following gibberish: id select_type table type possible_keys key key_len ref rows Extra 1 SIMPLE userdata ALL pageid 372420 1 SIMPLE page eq_ref PRIMARY,pagegroupid PRIMARY 4 topsecret.userdata.pageid 1 1 SIMPLE pagegroup eq_ref PRIMARY PRIMARY 4 topsecret.page.pagegroupid 1 Edit #2 SELECT u.field2, p.pageid FROM userdata u INNER JOIN page p ON u.pageid = p.pageid; /* 0.07 sec execution, 6.05 sec fecth */ id select_type table type possible_keys key key_len ref rows Extra 1 SIMPLE u ALL pageid 372420 1 SIMPLE p eq_ref PRIMARY PRIMARY 4 topsecret.u.pageid 1 Using index SELECT p.pageid, g.pagegroupid FROM page p INNER JOIN pagegroup g ON p.pagegroupid = g.pagegroupid; /* 9.37 sec execution, 60.0 sec fetch */ id select_type table type possible_keys key key_len ref rows Extra 1 SIMPLE g index PRIMARY PRIMARY 4 3646 Using index 1 SIMPLE p ref pagegroupid pagegroupid 5 topsecret.g.pagegroupid 3 Using where Moral of the story Keep medium/long text columns in a separate table if you run into performance problems such as this one.

    Read the article

  • Performance of VIEW vs. SQL statement

    - by Matt W.
    I have a query that goes something like the following: select <field list> from <table list> where <join conditions> and <condition list> and PrimaryKey in (select PrimaryKey from <table list> where <join list> and <condition list>) and PrimaryKey not in (select PrimaryKey from <table list> where <join list> and <condition list>) The sub-select queries both have multiple sub-select queries of their own that I'm not showing so as not to clutter the statement. One of the developers on my team thinks a view would be better. I disagree in that the SQL statement uses variables passed in by the program (based on the user's login Id). Are there any hard and fast rules on when a view should be used vs. using a SQL statement? What kind of performance gain issues are there in running SQL statements on their own against regular tables vs. against views. (Note that all the joins / where conditions are against indexed columns, so that shouldn't be an issue.) EDIT for clarification... Here's the query I'm working with: select obj_id from object where obj_id in( (select distinct(sec_id) from security where sec_type_id = 494 and ( (sec_usergroup_id = 3278 and sec_usergroup_type_id = 230) or (sec_usergroup_id in (select ug_gi_id from user_group where ug_ui_id = 3278) and sec_usergroup_type_id = 231) ) and sec_obj_id in ( select obj_id from object where obj_ot_id in (select of_ot_id from obj_form left outer join obj_type on ot_id = of_ot_id where ot_app_id = 87 and of_id in (select sec_obj_id from security where sec_type_id = 493 and ( (sec_usergroup_id = 3278 and sec_usergroup_type_id = 230) or (sec_usergroup_id in (select ug_gi_id from user_group where ug_ui_id = 3278) and sec_usergroup_type_id = 231) ) ) and of_usage_type_id = 131 ) ) ) ) or (obj_ot_id in (select of_ot_id from obj_form left outer join obj_type on ot_id = of_ot_id where ot_app_id = 87 and of_id in (select sec_obj_id from security where sec_type_id = 493 and ( (sec_usergroup_id = 3278 and sec_usergroup_type_id = 230) or (sec_usergroup_id in (select ug_gi_id from user_group where ug_ui_id = 3278) and sec_usergroup_type_id = 231) ) ) and of_usage_type_id = 131 ) and obj_id not in (select sec_obj_id from security where sec_type_id = 494) )

    Read the article

  • Improving performance in this query

    - by Luiz Gustavo F. Gama
    I have 3 tables with user logins: sis_login = administrators tb_rb_estrutura = coordinators tb_usuario = clients I created a VIEW to unite all these users by separating them by levels, as follows: create view `login_names` as select `n1`.`cod_login` as `id`, '1' as `level`, `n1`.`nom_user` as `name` from `dados`.`sis_login` `n1` union all select `n2`.`id` as `id`, '2' as `level`, `n2`.`nom_funcionario` as `name` from `tb_rb_estrutura` `n2` union all select `n3`.`cod_usuario` as `id`, '3' as `level`, `n3`.`dsc_nome` as `name` from `tb_usuario` `n3`; So, can occur up to three ids repeated for different users, which is why I separated by levels. This VIEW is just to return me user name, according to his id and level. considering it has about 500,000 registered users, this view takes about 1 second to load. too much time, but is becomes very small when I need to return the latest posts on the forum of my website. The tables of the forums return the user id and level, then look for a name in this VIEW. I have registered 18 forums. When I run the query, it takes one second for each forum = 18 seconds. OMG. This page loads every time somebody enter my website. This is my query: select `x`.`forum_id`, `x`.`topic_id`, `l`.`nome` from ( select `t`.`forum_id`, `t`.`topic_id`, `t`.`data`, `t`.`user_id`, `t`.`user_level` from `tb_forum_topics` `t` union all select `a`.`forum_id`, `a`.`topic_id`, `a`.`data`, `a`.`user_id`, `a`.`user_level` from `tb_forum_answers` `a` ) `x` left outer join `login_names` `l` on `l`.`id` = `x`.`user_id` and `l`.`level` = `x`.`user_level` group by `x`.`forum_id` asc USING EXPLAIN: id select_type table type possible_keys key key_len ref rows Extra 1 PRIMARY <derived2> ALL NULL NULL NULL NULL 6 Using temporary; Using filesort 1 PRIMARY <derived4> ALL NULL NULL NULL NULL 530415 4 DERIVED n1 ALL NULL NULL NULL NULL 114 5 UNION n2 ALL NULL NULL NULL NULL 2 6 UNION n3 ALL NULL NULL NULL NULL 530299 NULL UNION RESULT ALL NULL NULL NULL NULL NULL 2 DERIVED t ALL NULL NULL NULL NULL 3 3 UNION r ALL NULL NULL NULL NULL 3 NULL UNION RESULT ALL NULL NULL NULL NULL NULL Somebody can help me or give a suggestion?

    Read the article

  • Why do Scala maps have poor performance relative to Java?

    - by Mike Hanafey
    I am working on a Scala app that consumes large amounts of CPU time, so performance matters. The prototype of the system was written in Python, and performance was unacceptable. The application does a lot with inserting and manipulating data in maps. Rex Kerr's Thyme was used to look at the performance of updating and retrieving data from maps. Basically "n" random Ints were stored in maps, and retrieved from the maps, with the time relative to java.util.HashMap used as a reference. The full results for a range of "n" are here. Sample (n=100,000) performance relative to java, smaller is worse: Update Read Mutable 16.06% 76.51% Immutable 31.30% 20.68% I do not understand why the scala immutable map beats the scala mutable map in update performance. Using the sizeHint on the mutable map does not help (it appears to be ignored in the tested implementation, 2.10.3). Even more surprisingly the immutable read performance is worse than the mutable read performance, more significantly so with larger maps. The update performance of the scala mutable map is surprisingly bad, relative to both scala immutable and plain Java. What is the explanation?

    Read the article

  • Performance when querying a View

    - by Nate Bross
    I'm wondering if this is a bad practice or if in general this is the correct approach. Lets say that I've created a view that combines a few attributes from a few tables. My question, what do I need to do so I can query against this view as if it were a table without worrying about performance? All attributes in the original tables are indexed, my concern is that the result view will have hundreds of thousands of records, which I will want to narrow down quite a bit based on user input. What I'd like to avoid, is having multiple versions of the code that generates this view floating around with a few extra "where" conditions to facilitate the user input filtering. For example, assume my view has this header VIEW(Name, Type, DateEntered) this may have 100,000+ rows (possibly millions). I'd like to be able to make this view in SQL Server, and then in my application write querlies like this: SELECT Name, Type, DateEntered FROM MyView WHERE DateEntered BETWEEN @date1 and @date2; Basically, I am denormalizing my data for a series of reports that need to be run, and I'd like to centralize where I pull the data from, maybe I'm not looking at this problem from the right angle though, so I'm open to alternative ways to attack this.

    Read the article

  • Limiting object allocation over multiple threads

    - by John
    I have an application which retrieves and caches the results of a clients query. The client then requests different chunks of data and the application sends the relevant results and removes them from the cache. A new requirement for this application is that there needs to be a run-time configurable maximum number of results which may be cached. I've taken the naive approach and implemented this by using a counter under a lock which is incremented every time a result is cached and decremented whenever a result is removed from the cache. Unfortunately, this has drastically reduced the applications performance when processing a large number of concurrent requests. I have tried both a critical section lock and spin-lock; the performance improves a bit with a spin-lock, but is still unacceptably slow. Is there a better way to solve this problem which may improve performance? Right now I have a thread pool that services requests and each request is tied to a Request object which stores that cached results for that particular request. Here is a simplified pseudo code version of my current implementation: void ResultCallback( Result result, Request *request ) { lock totalResultsCached lock cachedLimit if( totalResultsCached + 1 > cachedLimit ) { unlock cachedLimit unlock totalResultsCached //cancel the request return; } ++totalResultsCached; unlock cachedLimit unlock totalResultsCached request.add(result) } void SendResults( int resultsToSend, Request *request ) { while ( resultsToSend > 0 ) { send(request.remove()) lock totalResultsCached --totalResultsCached unlock totalResultsCached --resultsToSend; } }

    Read the article

  • WPF Binding Failure performance hit vs Exception

    - by Aran Mulholland
    when we bind to heterogeneous collection of objects, not all objects have the same set of properties. in the output window we get a message like: System.Windows.Data Error: 39 : BindingExpression path error: 'RoundingFactor' property not found on 'object' ''MultiLineTextMarkingScheme' (HashCode=7262386)'. BindingExpression:Path=RoundingFactor;.......... This doesn't appear to be an exception, but we are concerned it has a performance impact. Should we be concerned and create a view model that has all the properties we wish to bind to (and have the properties that dont exist on the underlying element return null) or can we just leave it. This situation often occurs in a grid scenario where there might be large numbers of these binding failures.

    Read the article

  • HashSet vs. List performance

    - by Michael Damatov
    It's clear that a search performance of the generic HashSet<T> class is higher than of the generic List<T> class. Just compare the hash-based key with the linear approach in the List<T> class. However calculating a hash key may itself take some CPU cycles, so for a small amount of items the linear search can be a real alternative to the HashSet<T>. My question: where is the break-even? To simplify the scenario (and to be fair) let's assume that the List<T> class uses the element's Equals() method to identify an item.

    Read the article

  • Android Map Performance poor because of many Overlays?

    - by Dave
    Hi, I have a map in my android application that shows many markers (~20-50). But the app performs very poor when i try to scroll/zoom (in Google Android Maps i did a sample search for pizza and there were also some 20-50 results found and i didn't notice any particular performance problems when zooming/scrolling through the map). Here is my (pseudo)code: onCreate() { .... drawable = this.getResources().getDrawable(R.drawable.marker1); itemizedOverlay = new MyItemizedOverlay(drawable,mapView); ... callWebServiceToRetrieveData(); createMarkers(); } createMarkers(){ for(elem:bigList){ GeoPoint geoPoint = new GeoPoint((int)(elem.getLat()*1000000), (int) (elem.getLon()*1000000)); OverlayItem overlayItem = new OverlayItem(geoPoint, elem.getName(), elem.getData()); itemizedOverlay.addOverlay(overlayItem); mapOverlays.add(itemizedOverlay); } mapView.invalidate(); } the MyItemizedOverlay.addOverlay looks like this: public void addOverlay(OverlayItem overlay) { m_overlays.add(overlay); populate(); }

    Read the article

  • How to Avoid Your Next 12-Month Science Project

    - by constant
    While most customers immediately understand how the magic of Oracle's Hybrid Columnar Compression, intelligent storage servers and flash memory make Exadata uniquely powerful against home-grown database systems, some people think that Exalogic is nothing more than a bunch of x86 servers, a storage appliance and an InfiniBand (IB) network, built into a single rack. After all, isn't this exactly what the High Performance Computing (HPC) world has been doing for decades? On the surface, this may be true. And some people tried exactly that: They tried to put together their own version of Exalogic, but then they discover there's a lot more to building a system than buying hardware and assembling it together. IT is not Ikea. Why is that so? Could it be there's more going on behind the scenes than merely putting together a bunch of servers, a storage array and an InfiniBand network into a rack? Let's explore some of the special sauce that makes Exalogic unique and un-copyable, so you can save yourself from your next 6- to 12-month science project that distracts you from doing real work that adds value to your company. Engineering Systems is Hard Work! The backbone of Exalogic is its InfiniBand network: 4 times better bandwidth than even 10 Gigabit Ethernet, and only about a tenth of its latency. What a potential for increased scalability and throughput across the middleware and database layers! But InfiniBand is a beast that needs to be tamed: It is true that Exalogic uses a standard, open-source Open Fabrics Enterprise Distribution (OFED) InfiniBand driver stack. Unfortunately, this software has been developed by the HPC community with fastest speed in mind (which is good) but, despite the name, not many other enterprise-class requirements are included (which is less good). Here are some of the improvements that Oracle's InfiniBand development team had to add to the OFED stack to make it enterprise-ready, simply because typical HPC users didn't have the need to implement them: More than 100 bug fixes in the pieces that were not related to the Message Passing Interface Protocol (MPI), which is the protocol that HPC users use most of the time, but which is less useful in the enterprise. Performance optimizations and tuning across the whole IB stack: From Switches, Host Channel Adapters (HCAs) and drivers to low-level protocols, middleware and applications. Yes, even the standard HPC IB stack could be improved in terms of performance. Ethernet over IB (EoIB): Exalogic uses InfiniBand internally to reach high performance, but it needs to play nicely with datacenters around it. That's why Oracle added Ethernet over InfiniBand technology to it that allows for creating many virtual 10GBE adapters inside Exalogic's nodes that are aggregated and connected to Exalogic's IB gateway switches. While this is an open standard, it's up to the vendor to implement it. In this case, Oracle integrated the EoIB stack with Oracle's own IB to 10GBE gateway switches, and made it fully virtualized from the beginning. This means that Exalogic customers can completely rewire their server infrastructure inside the rack without having to physically pull or plug a single cable - a must-have for every cloud deployment. Anybody who wants to match this level of integration would need to add an InfiniBand switch development team to their project. Or just buy Oracle's gateway switches, which are conveniently shipped with a whole server infrastructure attached! IPv6 support for InfiniBand's Sockets Direct Protocol (SDP), Reliable Datagram Sockets (RDS), TCP/IP over IB (IPoIB) and EoIB protocols. Because no IPv6 = not very enterprise-class. HA capability for SDP. High Availability is not a big requirement for HPC, but for enterprise-class application servers it is. Every node in Exalogic's InfiniBand network is connected twice for redundancy. If any cable or port or HCA fails, there's always a replacement link ready to take over. This requires extra magic at the protocol level to work. So in addition to Weblogic's failover capabilities, Oracle implemented IB automatic path migration at the SDP level to avoid unnecessary failover operations at the middleware level. Security, for example spoof-protection. Another feature that is less important for traditional users of InfiniBand, but very important for enterprise customers. InfiniBand Partitioning and Quality-of-Service (QoS): One of the first questions we get from customers about Exalogic is: “How can we implement multi-tenancy?” The answer is to partition your IB network, which effectively creates many networks that work independently and that are protected at the lowest networking layer possible. In addition to that, QoS allows administrators to prioritize traffic flow in multi-tenancy environments so they can keep their service levels where it matters most. Resilient IB Fabric Management: InfiniBand is a self-managing network, so a lot of the magic lies in coming up with the right topology and in teaching the subnet manager how to properly discover and manage the network. Oracle's Infiniband switches come with pre-integrated, highly available fabric management with seamless integration into Oracle Enterprise Manager Ops Center. In short: Oracle elevated the OFED InfiniBand stack into an enterprise-class networking infrastructure. Many years and multiple teams of manpower went into the above improvements - this is something you can only get from Oracle, because no other InfiniBand vendor can give you these features across the whole stack! Exabus: Because it's not About the Size of Your Network, it's How You Use it! So let's assume that you somehow were able to get your hands on an enterprise-class IB driver stack. Or maybe you don't care and are just happy with the standard OFED one? Anyway, the next step is to actually leverage that InfiniBand performance. Here are the choices: Use traditional TCP/IP on top of the InfiniBand stack, Develop your own integration between your middleware and the lower-level (but faster) InfiniBand protocols. While more bandwidth is always a good thing, it's actually the low latency that enables superior performance for your applications when running on any networking infrastructure: The lower the latency, the faster the response travels through the network and the more transactions you can close per second. The reason why InfiniBand is such a low latency technology is that it gets rid of most if not all of your traditional networking protocol stack: Data is literally beamed from one region of RAM in one server into another region of RAM in another server with no kernel/drivers/UDP/TCP or other networking stack overhead involved! Which makes option 1 a no-go: Adding TCP/IP on top of InfiniBand is like adding training wheels to your racing bike. It may be ok in the beginning and for development, but it's not quite the performance IB was meant to deliver. Which only leaves option 2: Integrating your middleware with fast, low-level InfiniBand protocols. And this is what Exalogic's "Exabus" technology is all about. Here are a few Exabus features that help applications leverage the performance of InfiniBand in Exalogic: RDMA and SDP integration at the JDBC driver level (SDP), for Oracle Weblogic (SDP), Oracle Coherence (RDMA), Oracle Tuxedo (RDMA) and the new Oracle Traffic Director (RDMA) on Exalogic. Using these protocols, middleware can communicate a lot faster with each other and the Oracle database than by using standard networking protocols, Seamless Integration of Ethernet over InfiniBand from Exalogic's Gateway switches into the OS, Oracle Weblogic optimizations for handling massive amounts of parallel transactions. Because if you have an 8-lane Autobahn, you also need to improve your ramps so you can feed it with many cars in parallel. Integration of Weblogic with Oracle Exadata for faster performance, optimized session management and failover. As you see, “Exabus” is Oracle's word for describing all the InfiniBand enhancements Oracle put into Exalogic: OFED stack enhancements, protocols for faster IB access, and InfiniBand support and optimizations at the virtualization and middleware level. All working together to deliver the full potential of InfiniBand performance. Who else has 100% control over their middleware so they can develop their own low-level protocol integration with InfiniBand? Even if you take an open source approach, you're looking at years of development work to create, test and support a whole new networking technology in your middleware! The Extras: Less Hassle, More Productivity, Faster Time to Market And then there are the other advantages of Engineered Systems that are true for Exalogic the same as they are for every other Engineered System: One simple purchasing process: No headaches due to endless RFPs and no “Will X work with Y?” uncertainties. Everything has been engineered together: All kinds of bugs and problems have been already fixed at the design level that would have only manifested themselves after you have built the system from scratch. Everything is built, tested and integrated at the factory level . Less integration pain for you, faster time to market. Every Exalogic machine world-wide is identical to Oracle's own machines in the lab: Instant replication of any problems you may encounter, faster time to resolution. Simplified patching, management and operations. One throat to choke: Imagine finger-pointing hell for systems that have been put together using several different vendors. Oracle's Engineered Systems have a single phone number that customers can call to get their problems solved. For more business-centric values, read The Business Value of Engineered Systems. Conclusion: Buy Exalogic, or get ready for a 6-12 Month Science Project And here's the reason why it's not easy to "build your own Exalogic": There's a lot of work required to make such a system fly. In fact, anybody who is starting to "just put together a bunch of servers and an InfiniBand network" is really looking at a 6-12 month science project. And the outcome is likely to not be very enterprise-class. And it won't have Exalogic's performance either. Because building an Engineered System is literally rocket science: It takes a lot of time, effort, resources and many iterations of design/test/analyze/fix to build such a system. That's why InfiniBand has been reserved for HPC scientists for such a long time. And only Oracle can bring the power of InfiniBand in an enterprise-class, ready-to use, pre-integrated version to customers, without the develop/integrate/support pain. For more details, check the new Exalogic overview white paper which was updated only recently. P.S.: Thanks to my colleagues Ola, Paul, Don and Andy for helping me put together this article! var flattr_uid = '26528'; var flattr_tle = 'How to Avoid Your Next 12-Month Science Project'; var flattr_dsc = 'While most customers immediately understand how the magic of Oracle's Hybrid Columnar Compression, intelligent storage servers and flash memory make Exadata uniquely powerful against home-grown database systems, some people think that Exalogic is nothing more than a bunch of x86 servers, a storage appliance and an InfiniBand (IB) network, built into a single rack.After all, isn't this exactly what the High Performance Computing (HPC) world has been doing for decades?On the surface, this may be true. And some people tried exactly that: They tried to put together their own version of Exalogic, but then they discover there's a lot more to building a system than buying hardware and assembling it together. IT is not Ikea.Why is that so? Could it be there's more going on behind the scenes than merely putting together a bunch of servers, a storage array and an InfiniBand network into a rack? Let's explore some of the special sauce that makes Exalogic unique and un-copyable, so you can save yourself from your next 6- to 12-month science project that distracts you from doing real work that adds value to your company.'; var flattr_tag = 'Engineered Systems,Engineered Systems,Infiniband,Integration,latency,Oracle,performance'; var flattr_cat = 'text'; var flattr_url = 'http://constantin.glez.de/blog/2012/04/how-avoid-your-next-12-month-science-project'; var flattr_lng = 'en_GB'

    Read the article

  • Oracle Database 12c Spatial: Vector Performance Acceleration

    - by Okcan Yasin Saygili-Oracle
    Most business information has a location component, such as customer addresses, sales territories and physical assets. Businesses can take advantage of their geographic information by incorporating location analysis and intelligence into their information systems. This allows organizations to make better decisions, respond to customers more effectively, and reduce operational costs – increasing ROI and creating competitive advantage. Oracle Database, the industry’s most advanced database,  includes native location capabilities, fully integrated in the kernel, for fast, scalable, reliable and secure spatial and massive graph applications. It is a foundation for deploying enterprise-wide spatial information systems and locationenabled business applications. Developers can extend existing Oracle-based tools and applications, since they can easily incorporate location information directly in their applications, workflows, and services. Spatial Features The geospatial data features of Oracle Spatial and Graph option support complex geographic information systems (GIS) applications, enterprise applications and location services applications. Oracle Spatial and Graph option extends the spatial query and analysis features included in every edition of Oracle Database with the Oracle Locator feature, and provides a robust foundation for applications that require advanced spatial analysis and processing in the Oracle Database. It supports all major spatial data types and models, addressing challenging business-critical requirements from various industries, including transportation, utilities, energy, public sector, defense and commercial location intelligence. Network Data Model Graph Features The Network Data Model graph explicitly stores and maintains a persistent data model withnetwork connectivity and provides network analysis capability such as shortest path, nearest neighbors, within cost and reachability. It loads partitioned networks into memory on demand, overcomingthe limitations of in-memory analysis. Partitioning massive networks into manageable sub-networkssimplifies the network analysis. RDF Semantic Graph Features RDF Semantic Graph has native support for World Wide Web Consortium standards. It has open, scalable, and secure features for storing RDF/OWL ontologies anddata; native inference with OWL 2, SKOS and user-defined rules; and querying RDF/OWL data withSPARQL 1.1, Java APIs, and SPARQLgraph patterns in SQL. Video: Oracle Spatial and Graph Overview Oracle spatial is embeded on oracle database product. So ,we can use oracle installer (OUI).The Oracle Universal Installer (OUI) is used to install Oracle Database software. OUI is a graphical user interface utility that enables you to view the Oracle software that is installed on your machine, install new Oracle Database software, and delete Oracle software that you no longer need to use. Online Help is available to guide you through the installation process. One of the installation options is to create a database. If you select database creation, OUI automatically starts Oracle Database Configuration Assistant (DBCA) to guide you through the process of creating and configuring a database. If you do not create a database during installation, you must invoke DBCA after you have installed the software to create a database. You can also use DBCA to create additional databases. For installing Oracle Database 12c you may check the Installing Oracle Database Software and Creating a Database tutorial under the Oracle Database 12c 2-Day DBA Series.You can always check if spatial is available in your database using  "select comp_id, version, status, comp_name from dba_registry where comp_id='SDO';"   One of the most notable improvements with Oracle Spatial and Graph 12c can be seen in performance increases in vector data operations. Enabling the Spatial Vector Acceleration feature (available with the Spatial option) dramatically improves the performance of commonly used vector data operations, such as sdo_distance, sdo_aggr_union, and sdo_inside. With 12c, these operations also run more efficiently in parallel than in prior versions through the use of metadata caching. For organizations that have been facing processing limitations, these enhancements enable developers to make a small set of configuration changes and quickly realize significant performance improvements. Results include improved index performance, enhanced geometry engine performance, optimized secondary filter optimizations for Spatial operators, and improved CPU and memory utilization for many advanced vector functions. Vector performance acceleration is especially beneficial when using Oracle Exadata Database Machine and other large-scale systems. Oracle Spatial and Graph vector performance acceleration builds on general improvements available to all SDO_GEOMETRY operations in these areas: Caching of index metadata, Concurrent update mechanisms, and Optimized spatial predicate selectivity and cost functions. These optimizations enable more efficient use of: CPU, Memory, and Partitioning Resulting in substantial query performance improvements.UsageTo accelerate the performance of spatial operators, it is recommended that you set the SPATIAL_VECTOR_ACCELERATION database system parameter to the value TRUE. (This parameter is authorized for use only by licensed Oracle Spatial users, and its default value is FALSE.) You can set this parameter for the whole system or for a single session. To set the value for the whole system, do either of the following:Enter the following statement from a suitably privileged account:   ALTER SYSTEM SET SPATIAL_VECTOR_ACCELERATION = TRUE;Add the following to the database initialization file (xxxinit.ora):   SPATIAL_VECTOR_ACCELERATION = TRUE;To set the value for the current session, enter the following statement from a suitably privileged account:   ALTER SESSION SET SPATIAL_VECTOR_ACCELERATION = TRUE; Checkout the complete list of new features on Oracle.com @ http://www.oracle.com/technetwork/database/options/spatialandgraph/overview/index.html Spatial and Graph Data Sheet (PDF) Spatial and Graph White Paper (PDF)

    Read the article

  • ruby / rails / mysql performance degraded on Snow Leopard

    - by adamaig
    I've burned a bunch of hours on this. I'm not having problems getting things to build, but I am seeing that my test suite runs about 2x slower than when I was on OS X 10.5.x . I've spent a lot of time playing around with different optimization settings (learning to avoid homebrew's llvm-gcc compilation). I've just learned that I needed to tweaks /Library/Preferences/SystemConfiguration/com.apple.Boot.plist in order to get the kernel to boot in 64 bit mode. However, my rails app is still running a bit slower than before, even after warming up the mysql server. So what performance tweaks might i need to look into? Right now the stock ruby 1.8.7 runs faster than 1.9.1 for some things, and I'd really like to know if there is anything I should be looking for. All my dev software has been compiled for x86_64, mysql with -O2 optimization, using regular gcc (not llvm-gcc).

    Read the article

  • Rails 2.3.5 model name translation problem in error messages

    - by Jason Nerer
    Hi Rails'ers, I encountered some problem while trying to translate my model's names and attributes in a Rails 2.3.5 app. I have the following model: class BillingPlan < ActiveRecord::Base validates_presence_of :billing_option_id belongs_to :order belongs_to :user belongs_to :billing_option end When validation fails, my models attributes are translated correctly, but the modelname itself is not. I use the following translation skeleton in de.yml de: activerecord: models: shipping_plan: "Versandart" billing_plan: "Rechnungsart" attributes: shipping_plan: shipping_option_id: "Versandoption" billing_plan: billing_option_id: "Rechnungsoption" Basis for my translation file is: http://github.com/svenfuchs/rails-i18n/blob/master/rails/locale/de.yml Can anyone help? Thx in advance J.

    Read the article

  • Performance of DrawingVisual vs Canvas.OnRender for lots of constantly changing shapes

    - by romkyns
    I'm working on a game-like app which has up to a thousand shapes (ellipses and lines) that constantly change at 60fps. Having read an excellent article on rendering many moving shapes, I implemented this using a custom Canvas descendant that overrides OnRender to do the drawing via a DrawingContext. The performance is quite reasonable, although the CPU usage stays high. However, the article suggests that the most efficient approach for constantly moving shapes is to use lots of DrawingVisual instances instead of OnRender. Unfortunately though it doesn't explain why that should be faster for this scenario. Changing the implementation in this way is not a small effort, so I'd like to understand the reasons and whether they are applicable to me before deciding to make the switch. Why could the DrawingVisual approach result in lower CPU usage than the OnRender approach in this scenario?

    Read the article

  • JavaScript replace with callback - performance question

    - by Tomalak
    In JavaScript, you can define a callback handler in regex string replace operations: str.replace(/str[123]|etc/, replaceCallback); Imagine you have a lookup object of strings and replacements. var lookup = {"str1": "repl1", "str2": "repl2", "str3": "repl3", "etc": "etc" }; and this callback function: var replaceCallback = function(match) { if (lookup[match]) return lookup[match]; else return match; } How would you assess the performance of the above callback? Are there solid ways to improve it? Would if (match in lookup) //.... or even return lookup[match] | match; lead to opportunities for the JS compiler to optimize, or is it all the same thing?

    Read the article

  • Monitoring process-level performance counters in Windows Perfmon

    - by Dennis Kashkin
    I am sure everybody has bumped into this. As you scale a web server that uses multiple application pools, it's valuable to collect performance counters for each application pool 24x7. The only problem is - Perfmon links counters to application pools by process ID, so whenever an application pool recycles you have to remove the counters for the old process ID and add them for the new process ID. Since application pools recycle quite often (whenever you release a new version or patch the server), it's a major pain. I wonder if anybody has found a workaround for this? Perhaps a programmatic way to update Perfmon settings whenever an application pool starts up or some way to reference application pools by name instead of process ID? I'll appreciate any hints on this!

    Read the article

< Previous Page | 40 41 42 43 44 45 46 47 48 49 50 51  | Next Page >