Search Results

Search found 12398 results on 496 pages for 'in memory oltp'.

Page 285/496 | < Previous Page | 281 282 283 284 285 286 287 288 289 290 291 292  | Next Page >

  • C# LZO Library and Experience

    - by Hounshell
    Anyone have experience with a C# LZO compression/decompression library? LZO.NET (at http://lzo-net.sourceforge.net/ ) looks pretty alpha QuickLZ (at http://www.quicklz.com/ ) isn't stream-based and I need to compress files as they're generated and don't want to buffer the whole file in memory MiniLZO (at http://www.codeproject.com/KB/recipes/managedlzo.aspx ) is on CodeProject and I don't have a good track record with code from there working

    Read the article

  • SQLSTATE 08001 error

    - by Joseph
    Hi all OUR SQL Server 2008 got freeze.and not able to connect the databse and we restrted the server(windows 2008). Even we were not able to login windows(Restrted via HP ILO). ANybody face this issue?and have a resolution. [165] ODBC Error: 0, Login timeout expired [SQLSTATE HYT00] [298] SQLServer Error: 258, Shared Memory Provider: Timeout error [258]. [SQLSTATE 08001] [382] Logon to server '(local)' failed (ConnUpdateJobActivity_NextScheduledRunDate) the above three error we got this time thanks in advance

    Read the article

  • OperationalError "unable to open database file" processing query results with SQLAlchemy and SQLite3

    - by Peter
    I'm running into this little problem that I hope is just a dumb user error. It looks like some sort of a size limit with a query to a SQLite database. I managed to reproduce the issue with an in-memory DB and a simple script shown below. I can make it work by either reducing the number of records in the DB; or by reducing the size of each record; or by dropping the order_by() call. I am using Python 2.5.5 and SQLAlchemy 0.6.0 in a Cygwin environment. Thanks! #!/usr/bin/python from sqlalchemy.orm import sessionmaker import sqlalchemy import sqlalchemy.orm class Person(object): def __init__(self, name): self.name = name engine = sqlalchemy.create_engine('sqlite:///:memory:') Session = sessionmaker(bind=engine) metadata = sqlalchemy.schema.MetaData(bind=engine) person_table = sqlalchemy.Table('person', metadata, sqlalchemy.Column('id', sqlalchemy.types.Integer, primary_key=True), sqlalchemy.Column('name', sqlalchemy.types.String)) metadata.create_all(engine) sqlalchemy.orm.mapper(Person, person_table) session = Session() session.add_all([Person("012345678901234567890123456789012") for i in range(5000)]) session.commit() persons = session.query(Person).order_by(Person.name).all() print "count =", len(persons) session.close() The all() call to the query result fails with the OperationalError exception: Traceback (most recent call last): File "./stress.py", line 27, in <module> persons = session.query(Person).order_by(Person.name).all() File "/usr/lib/python2.5/site-packages/sqlalchemy/orm/query.py", line 1343, in all return list(self) File "/usr/lib/python2.5/site-packages/sqlalchemy/orm/query.py", line 1451, in __iter__ return self._execute_and_instances(context) File "/usr/lib/python2.5/site-packages/sqlalchemy/orm/query.py", line 1456, in _execute_and_instances mapper=self._mapper_zero_or_none()) File "/usr/lib/python2.5/site-packages/sqlalchemy/orm/session.py", line 737, in execute clause, params or {}) File "/usr/lib/python2.5/site-packages/sqlalchemy/engine/base.py", line 1109, in execute return Connection.executors[c](self, object, multiparams, params) File "/usr/lib/python2.5/site-packages/sqlalchemy/engine/base.py", line 1186, in _execute_clauseelement return self.__execute_context(context) File "/usr/lib/python2.5/site-packages/sqlalchemy/engine/base.py", line 1215, in __execute_context context.parameters[0], context=context) File "/usr/lib/python2.5/site-packages/sqlalchemy/engine/base.py", line 1284, in _cursor_execute self._handle_dbapi_exception(e, statement, parameters, cursor, context) File "/usr/lib/python2.5/site-packages/sqlalchemy/engine/base.py", line 1282, in _cursor_execute self.dialect.do_execute(cursor, statement, parameters, context=context) File "/usr/lib/python2.5/site-packages/sqlalchemy/engine/default.py", line 277, in do_execute cursor.execute(statement, parameters) sqlalchemy.exc.OperationalError: (OperationalError) unable to open database file u'SELECT person.id AS person_id, person.name AS person_name \nFROM person ORDER BY person.name' ()

    Read the article

  • c# write big files to blob sqlite

    - by brizjin-gmail-com
    I have c# application which write files to sqlite database. It uses entity fraemwork for modeling data. Write file to blob (entity byte[] varible) with this line: row.file = System.IO.File.ReadAllBytes(file_to_load.FileName); //row.file is type byte[] //row is entity class table All work correctly when files size is less. When size more 300Mb app throw exception: Exception of type 'System.OutOfMemoryException' was thrown. How I can write to blob direct, without memory varibles?

    Read the article

  • how caching is implemented using PHP GD

    - by christian
    I want to cache the images of my gallery. Generating images every pages load using GD uses a lot of memory, So I am planning to create a cache image for the images generated by php script done with GD. What will be the best thing to create cache?

    Read the article

  • How to build OpenSSL in release mode with symbols?

    - by twk
    I generally build my openssl libs by doing this: perl Configure VC-WIN32 ms\do_masm nmake -f ms\ntdll.mak nmake -f ms\ntdll.mak test For debugging minidumps, I'd like to get a pdb file for the 2 dlls (while still building them in "release" mode). I've added /Zi to the CFLAGS part of ms\ntdll.mak, but I still get "Binary was not built with debug information" when I look at the in-memory modules in the visual studio debugger. Am I missing a step? Thanks

    Read the article

  • Restoring/Restarting a java daemon from crash

    - by Ashish
    Hello, I am running a java app as daemon on a linux machine using a customized shell script. Since I am new to java and linux, I want to know is it possible that the app itself resurrects itself(just like restart) and recovers from cases like app crashing, unhandled exceptions or out of memory etc. thanks in advance Ashish Sharma

    Read the article

  • Is dealloc method called in presentModalViewController when dismiss the controller.

    - by Madan Mohan
    Hi Guys, here the following code is used to view the present modal view controller. [[self navigationController] presentModalViewController:doctorListViewNavigationController animated:YES]; code for dismiss the ModalViewController. -(void)closeAction { [[self navigationController] dismissModalViewControllerAnimated:YES]; } My problem is the dealloc method is not called then I am getting memory issue problems like object allocations, leaks.. (void)dealloc { [doctorList release]; [myTableView release]; [super dealloc]; }

    Read the article

  • How to update a user created Bitmap in the Windows API

    - by gamernb
    In my code I quickly generate images on the fly, and I want to display them as quickly as possible. So the first time I create my image, I create a new BITMAP, but instead of deleting the old one and creating a new one for every subsequent image, I just want to copy my data back into the existing one. Here is my code to do both the initial creation and the updating. The creation works just fine, but the updating one doesn't work. BITMAPINFO bi; HBITMAP Frame::CreateBitmap(HWND hwnd, int tol1, int tol2, bool useWhite, bool useBackground) { ZeroMemory(&bi.bmiHeader, sizeof(BITMAPINFOHEADER)); bi.bmiHeader.biSize = sizeof(BITMAPINFOHEADER); bi.bmiHeader.biWidth = width; bi.bmiHeader.biHeight = height; bi.bmiHeader.biPlanes = 1; bi.bmiHeader.biBitCount = 24; bi.bmiHeader.biCompression = BI_RGB; ZeroMemory(bi.bmiColors, sizeof(RGBQUAD)); // Allocate memory for bitmap bits int size = height * width; Pixel* newPixels = new Pixel[size]; // Recompute the output //memcpy(newPixels, pixels, size*3); ComputeOutput(newPixels, tol1, tol2, useWhite, useBackground); HBITMAP bitmap = CreateDIBitmap(GetDC(hwnd), &bi.bmiHeader, CBM_INIT, newPixels, &bi, DIB_RGB_COLORS); delete newPixels; return bitmap; } and void Frame::UpdateBitmap(HWND hwnd, HBITMAP bitmap, int tol1, int tol2, bool useWhite, bool useBackground) { ZeroMemory(&bi.bmiHeader, sizeof(BITMAPINFOHEADER)); bi.bmiHeader.biSize = sizeof(BITMAPINFOHEADER); HDC hdc = GetDC(hwnd); if(!GetDIBits(hdc, bitmap, 0, bi.bmiHeader.biHeight, NULL, &bi, DIB_RGB_COLORS)) MessageBox(NULL, "Can't get base image info!", "Error!", MB_ICONEXCLAMATION | MB_OK); // Allocate memory for bitmap bits int size = height * width; Pixel* newPixels = new Pixel[size]; // Recompute the output //memcpy(newPixels, pixels, size*3); ComputeOutput(newPixels, tol1, tol2, useWhite, useBackground); // Push back to windows if(!SetDIBits(hdc, bitmap, 0, bi.bmiHeader.biHeight, newPixels, &bi, DIB_RGB_COLORS)) MessageBox(NULL, "Can't set pixel data!", "Error!", MB_ICONEXCLAMATION | MB_OK); delete newPixels; } where the Pixel struct is just this: struct Pixel { unsigned char b, g, r; }; Why does my update function not work. I always get the MessageBox for "Can't set pixel data!" I used code similar to this when I was loading in the original bitmap from file, then editing the data, but now when I manually create it, it doesn't work.

    Read the article

  • Huge anon blocks in pmap

    - by Parik
    I am doing a pmap on a tomcat process and I am seeing some huge anon blocks. From what I read anon blocks are used for thread stacks and for JNI. I have a very moderate thread count. How can I go about finding out what is causing these huge anon blocks? 00000000ee0d0000 26752K rwx-- [ anon ] 00000000efaf0000 33792K rwx-- [ anon ] 00000000f1bf0000 25856K rwx-- [ anon ] 00000000f3530000 39680K rwx-- [ anon ] ( on a side note is pmap the correct way to measure how much memory is allocated to tomcat?)

    Read the article

  • Java OutOfMemory error in filter

    - by The Machine
    I get a java.lang.outOfMemoryError exception, while writing a big file to the servletOutputStream. Every response is by default wrapped using a ehcache.constructs.web.filter Object for GZIP compression. And as per the logs, the exception is thrown in the Filter object. Is there a way to increase the available memory so, that the outOfMemoryError Exception does not occur ?

    Read the article

  • does Entity Framework have a caching mechanism like DataSet/DataTables have?

    - by Greg
    Hi, does Entity Framework have a caching mechanism like DataSet/DataTables have? In other words if you start using EF is it really just a way to easily get data in/out of the database without providing an additional caching layer like DataSet/DataTables did where you have an in-memory representation (on-line/off-line) and which at some point you could say to "persist all changes to database"

    Read the article

  • Java without gc - io

    - by Dan
    Hi Guys I would like to run a Java program with garbage collection switched off. Managing memory in my own code is not so difficult. However the program needs quite a lot of I/O. Is there any way (short of using JNI for all I/O operations) that I could achieve this using pure Java? Thanks Daniel

    Read the article

  • Stream multiple files in _one_ ASIHTTPRequest

    - by Snej
    What is best practice to stream multiple files in one ASIHTTPRequest? Right now, for one file I use: .... ASIHTTPRequest *request = [[ASIHTTPRequest alloc] initWithURL:someUrl]; [request setShouldStreamPostDataFromDisk:YES]; [request appendPostDataFromFile:someFilePath]; [request startSynchronous]; How to stream multiple files without placing all files in memory before transmission?

    Read the article

  • Is redis a durable datastore?

    - by allyourcode
    By "durable" I mean, the server can crash at any time, and as long as the disk remains in tact, no data is lost (see ACID). Seems like that's what journaling mode is for, but if you enable journaling, doesn't that defeat the purpose of operating on in-memory data? Read operations might not be affected by journaling, but it seems like journaling would kill your write performance.

    Read the article

  • Uploading Large Files

    - by mohamed hadad
    Hello, I'm using a windows service on my server to receive large files (1 GB) from desktop clients when I use the TCPClient class it creates astream to send the file which bloks my memory, What is the well designe for this problem

    Read the article

< Previous Page | 281 282 283 284 285 286 287 288 289 290 291 292  | Next Page >