Search Results

Search found 10028 results on 402 pages for 'berkeley db'.

Page 91/402 | < Previous Page | 87 88 89 90 91 92 93 94 95 96 97 98  | Next Page >

  • How to see active connections and "current activity" in PostgreSQL 8.4

    - by Evgeny
    I'm investigating an issue with DB connections being left open indefinitely, causing problems on the DB server. How do I see currently open connections to a PostgreSQL server, particularly those using a specific database? Ideally I'd like to see what command is executing there as well. Basically, I'm looking for something equivalent to the "Current Activity" view in MSSQL.

    Read the article

  • Mysql charset problem

    - by Newtonx
    I'm trying to import some data from one server to another. But when I do it, I'm having problems with charset. Words like Goiânia became Goiâni and conceição became conceição My Application was set to use latin1 charset Server 1 : MySQL Charset : UTF-8 Unicode (utf8) table collation : latin1_swedish_ci Server 2 : MySQL Charset: UTF-8 Unicode (utf8) table collation : latin1_swedish_ci Command I used to export data from server 1 mysqldump -u root -p --default-character-set=iso-8859-1 database_name db.sql Command used to restore to server 2 mysql -u root -p database_name < db.sql

    Read the article

  • Just LB or also Web Servers in Demilitarized Zone?

    - by Bradford
    In a load balanced environment, is it necessary to have all of the web servers in the DMZ? Or will just having the Load Balancer in the DMZ achieve the desired security? If it matters, the web server and application server are the same -- GF, Tomcat fronted by httpd on the same server, OAS, etc... LB - WEB/APPLICATION - DB Also, would the setup be different if it was LB - Web Server - Application Server - DB Thanks, Bradford

    Read the article

  • PostgreSQL encrypted backups

    - by Nikhil Gupte
    Is it possible to ensure that dumps taken from a PostgreSQL db are always encrypted? The data in the database is highly sensitive and we cannot afford un-authorized personnel, including Sys Admins who need to backup the db, to access the actual data.

    Read the article

  • PostgreSQL Backup/Restore

    - by Ian
    What's the best way to backup a postgresql database? I've tried using the documentation at www.postgresql.org, but I always get integrity errors when restoring. Right now I'm using this for backup: pg_dump -U myuser -d mydatabase db.pg.dump for restore: pg_restore -c -r -U myuser -d mydatabase db.pg.dump But I'm not getting the desired results..

    Read the article

  • What the best way to achieve RPO of zero and lowest possible RTO (less than 15 minutes) with SQL 2008 R2?

    - by Adrian Hope-Bailie
    We are running a payments (EFT transaction processing) application which is processing high volumes of transactions 24/7 and are currently investigating a better way of doing DB replication to our disaster recovery site. Our current and previous strategies have included using both DoubleTake and Redgate to replicate data to a warm stand-by. DoubleTake is the supported solution from the payments software vendor however their (DoubleTake's) support in South Africa is very poor. We had a few issues and simply couldn't ever resolve them so we had to give up on DoubleTake. We have been using Redgate to manually read the data from the primary site (via queries) and write to the DR site but this is: A bad solution Getting the software vendor hot and bothered whenever we have support issues as it has a tendency to interfere with the payment application which is very DB intensive. We recently upgraded the whole system to run on SQL 2008 R2 Enterprise which means we should probably be looking at using some of the built-in replication features. The server has 2 fairly large databases with a mixture of tables containing highly volatile transactional data and pretty static configuration data. Replication would be done over a WAN link to a separate physical site and needs to achieve the following objectives. RPO: Zero loss - This is transactional data with financial impact so we can't lose anything. RTO: Tending to zero - The business depends on our ability to process transactions every minute we are down we are losing money I have looked at a few of the other questions/answers but none meet our case exactly: SQL Server 2008 failover strategy - Log shipping or replication? How to achieve the following RTO & RPO with logshipping only using SQL Server? What is the best of two approaches to achieve DB Replication? My current thinking is that we should use mirroring but I am concerned that for RPO:0 we will need to do delayed commits and this could impact the performance of the primary DB which is not an option. Our current DR process is to: Stop incoming traffic to the primary site and allow all in-flight transaction to complete. Allow the replication to DR to complete. Change network routing to route to DR site. Start all applications and services on the secondary site (Ideally we can change this to a warmer stand-by whereby the applications are already running but not processing any transactions). In other words the DR database needs to, as quickly as possible, catch up with primary and be ready for processing as the new primary. We would then need to be able to reverse this when we are ready to switch back. Is there a better option than mirroring (should we be doing log-shipping too) and can anyone suggest other considerations that we should keep in mind?

    Read the article

  • Accessing Internal Server from Outside through SSH

    - by myahya
    The issue is as follows: I am using Ubuntu Linux. I am working from home, and I need to access a PostgreSQL database severer within my company's network. Call the DB server X. For security reasons, only one machine can be accessed from the outside using ssh, call it Y. So the scenario is: Home <--SSH-- Y <----- X(DB) I think something like SSH forwarding (configured in .ssh/config) could help with this, but I am not sure how to do this, any ideas?

    Read the article

  • How to get the MongoDB' current working set size

    - by Howard
    From the doc , it said "For best performance, the majority of your active set should fit in RAM." So for example, my db.stats() give me { "db" : "mydb", "collections" : 16, "objects" : 21452, "avgObjSize" : 768.0516501957859, "dataSize" : 16476244, "storageSize" : 25385984, "numExtents" : 43, "indexes" : 70, "indexSize" : 15450112, "fileSize" : 469762048, "ok" : 1 } Which value is the working set size?

    Read the article

  • Postgresql Internals - Documentation

    - by NogginTheNog
    I'm looking for some up-to-date information about postgresql internals, specifically the query optimizer. I've found this link (referred to in the "Further Reading" section of the 8.4 docs):- http://db.cs.berkeley.edu//papers/UCB-MS-zfong.pdf but it seems quite old. That in itself is not a problem, but I wanted to be sure that I have information that is relevant. Is this the best resource for understanding how postgresql processes queries (using plans, statistics etc.) or are there others?

    Read the article

  • What is the best powershell script to restore an SQL Database?

    - by EtienneT
    To restore an SQL Server 2008 database, I would lile to be able to just do something like this in powershell: ./restore.ps1 DatabaseName.bak Then the powershell script would by convention restore it to a database with name "DatabaseName". It would disconnect any user connected to this database so that it can restore the DB. It would store the mdf and ldf in the default location. This would mainly be while developing on my personal machine. Just a quick way to restore a DB. Anyone has such a script? Thanks

    Read the article

  • Subversion hard disk full

    - by user37914
    Our SVN server hd was full. Deleted a load of unwanted data on there, and all appeared to be well. However, new check-ins are now failing. Can't move '/repos/db/txn-protorevs/1000-vr.rev' to '/repos/db/revs/1/1001': Permission denied 1000-vr.rev does not exist. There are two other files in there: 9-9.rev and 9-9.rev-lock Edit: I have checked file permissions just in case, and all is as it should be.

    Read the article

  • Need help in using svn on ubuntu 9.10

    - by michael
    Hi, I have install svn on ubuntu 9.10. But when I try to use svn to checkout code for an open source project, i get this error: $ svn co svn://svn.valgrind.org/valgrind/trunk valgrind svn: Berkeley DB error for filesystem '/home/svn/repos/valgrind/db' while opening 'nodes' table: Cannot allocate memory svn: bdb: Lock table is out of available locker entries Can you please tell me how to fix it? Thank you.

    Read the article

  • SQL Server 2008 to SQL Server 2008

    - by Sakhawat Ali
    I have an MDF and LDF file of SQL Server 2005. i attached it with SQL Server 2008 and did some change in data. now when i attached it back to sql server 2005 Express Edition it gives version error. The database 'E:\DB\JOBPERS.MDF' cannot be opened because it is version 655. This server supports version 612 and earlier. A downgrade path is not supported. Could not open new database 'E:\DB\JOBPERS.MDF'. CREATE DATABASE is aborted. An attempt to attach an auto-named database for file E:\DB\Jobpers.mdf failed. A database with the same name exists, or specified file cannot be opened, or it is located on UNC share.

    Read the article

  • Populate Android Database From CSV file?

    - by MoMo
    Is it possible to take a csv file stored in the res/raw resource directory and use it to populate a table in the sqlite3 database? My thought was that, if there was a way to do a bulk import for the entire file into the table then that would be cleaner and faster than iterating over each line in the file and executing individual insert statements... I've found that there is a sqlite import command that allows this: http://stackoverflow.com/questions/1045910/how-can-i-import-load-a-sql-or-csv-file-into-sqlite ...but I'm having trouble applying those statements in my Android application. My first thought was to try something like the following...but no luck: db.execSQL("CREATE TABLE " + TABLE_NAME + "(id INTEGER PRIMARY KEY, name TEXT)"); db.execSQL(".mode csv"); db.execSQL(".import res/raw/MyFile.csv " + TABLE_NAME); Is this possible? Should I be trying a different approach to populate my database? Thanks for you answers!

    Read the article

  • Retrieving inline images from Lotus notes using lotusscript

    - by Nazrul
    I have some NotesDocument where some RichText fields have both text and inline images. I can get text part of that items but can't retrieve inline images using lotusscript. Could any one please suggest me a way to retrieve inline images from that documents. LotusScript code: Sub Click(Source As Button) Dim session As New NotesSession Dim db As NotesDatabase Dim mainDoc As NotesDocument Dim v As NotesView Set db = session.CurrentDatabase Dim fileName As String Dim fileNum As Integer fileNum% = Freefile() fileName$ = "D:\data.txt" Open FileName$ For Append As fileNum% Set v = db.GetView("MyView") Set mainDoc = v.GetFirstDocument While Not ( mainDoc Is Nothing ) Forall i In mainDoc.Items If i.Type = RICHTEXT Then Write #fileNum% , i.Name & ":" & i.text 'how the images?? End If End Forall Set mainDoc = v.GetNextDocument( mainDoc ) Wend End Sub Thanks.

    Read the article

  • How to group data in PHP inside the while loop

    - by christian
    I query in a database and loop the result using while and heres the sample data retrieve: echo $db->f("FirstName")."===".$db->f("Question")."=".$db->f("Answer")."<br>"; Michael===Q2=allergy Michael===Q2=Hives Michael===Q6=A lot Michael===Q8_A=Daktacort Michael===Q1=Itch Michael===Q5=Smoke Michael===Q8_A=Diprogenta Christian===Q1=Stuffy Christian===Q6=A lot Christian===Q1=Clear Christian===Q5=Pollen How can I group them according to name and to their Q value? I want something like this result: Name Q1 Q2 ..... so on and so fort. Michael Itch Hives, Allergy Christian Stuffy

    Read the article

  • PyQt error: No such signal QObject::dataChanged

    - by DSblizzard
    PyQt application works fine, but when I close it Python shows this message: "Object::connect: No such signal QObject::dataChanged(QModelIndex,QModelIndex)" What is the cause of this? There isn't dataChanged signal in the program. EDIT: Almost minimal program which causes error: import sys from PyQt4.QtCore import * from PyQt4.QtGui import * from PyQt4.QtSql import * import ui_DBMainWindow global Mw, Table, Tv Id, Name, Size = range(3) class TTable(): pass Table = TTable() class TMainWindow(QMainWindow, ui_DBMainWindow.Ui_MainWindow): def __init__(self, parent = None): global Table QMainWindow.__init__(self, parent) self.setupUi(self) self.showMaximized() self.mapper = QDataWidgetMapper(self) self.mapper.setModel(Table.Model) def main(): global Mw, Table, Tv QApp = QApplication(sys.argv) DB = QSqlDatabase.addDatabase("QSQLITE") DB.setDatabaseName("1.db") Table.Model = QSqlTableModel() Table.Model.setTable("MainTable") Table.Model.select() Mw = TMainWindow() QApp.exec_() if __name__ == "__main__": main()

    Read the article

  • Schema compare with MS Data Tools in VS2008

    - by rdkleine
    When performing a schema compare having db_owner rights on the target database results in the following error: The user does not have permission to perform this action. Using the SQL Server Profiler I figured out this error occurs executing a query targeting the master db view: [sys].[dm_database_encryption_keys] While specifically ignoring all object types but Tables one would presume the SQL Compare doesn't need access to the db encryption keys. Also note: http://social.msdn.microsoft.com/Forums/en-US/vstsdb/thread/c11a5f8a-b9cc-454f-ba77-e1c69141d64b/ One solution would be to GRANT VIEW SERVER STATE to the db user, but in my case I'm not hosting the database services and won't get the rights to the server state. Also tried excluding DatabaseEncryptionKey element in the compare file. <PropertyElementName> <Name>Microsoft.Data.Schema.Sql.SchemaModel.SqlServer.ISql100DatabaseEncryptionKey</Name> <Value>ExcludedType</Value> </PropertyElementName> Anyone has an workaround this?

    Read the article

  • Search french via Thinking Sphinx in ROR project

    - by Daniel
    Hello everyone: How can I use sphinx to search french words which the entries in the db is actually english? The situation is: I have a ROR project with a table in the db called "categories", and the category names are in english, category has many "question" entries. In localization file config/locals/fr.yml, these categories were translated to french. Consider about expandability, we can't change the category names in the db to french. User can search by type part of the key word. Here is a example: Category Name: Health and Medical In french: Santé et médecine so how can I do this: type "Santé médecine abc" in the search field and sphinx will return the "questions" under "Health and Medical" category and have keyword "abc"?

    Read the article

  • Android: Filtering a SimpleCursorAdapter ListView

    - by Diego Tori
    Right now, I'm running into issues trying to implement a FilterQueryProvider in my custom SimpleCursorAdapter, since I'm unsure of what to do in the FilterQueryProvider's runQuery function. In other words, since the query that comprises my ListView basically gets the rowID, name, and a third column from my databases's table, I want to be able to filter the cursor based on the partial value of the name column. However, I am uncertain of whether I can do this directly from runQuery without expanding my DB class since I want to filter the existing cursor, or will I have to create a new query function in my DB class that partially searches my name column, and if so, how would I go about creating the query statement while using the CharSequence constraint argument in runQuery? I am also concerned about the performance issues associated with trying to run multiple queries based on partial text since the DB table in question has about 1300-1400 rows. In other words, would I run into a bottleneck trying to filter the cursor?

    Read the article

  • Strange problem with MySQL full text query

    - by Manish
    This probably has something to do with my understanding of full text search or perhaps a known issue. My Java code creates search term for full text search on MySQL like this - +word* This works fine if the value in the DB column contains more text than the word itself. However, if the value is exact - no result are returned. I expected it to return results when value is an exact match. Let me give an example - Assume that DB has column name with value "Manish", now if I search with this - +Manis* It does return the result correctly, but if I try this - +Manish* It doesn't return any result, though exact match exists in DB column - Name. How can I obtain the desired behaviour in both these cases? Removing + sign works, but it returns far too many results when there are two or more words. Any help, pointers would be highly appreciated! I am using MySQL 5.0 TIA, - Manish

    Read the article

  • What's the convention for extending Linq with set based helper operations

    - by Luke Rohde
    Hi All I might be vaguing out here but I'm looking for a nice place to put set based helper operations in linq so I can do things like; db.Selections.ClearTemporary() which does something like db.DeleteAllOnSubmit(db.Selections.Where(s => s.Temporary)) Since I can figure out how to extend Table<Selection> the best I can do is create a static method in partial class of Selection (similar to Ruby) but I have to pass in the datacontext like; Selection.ClearTemporary(MyDataContext) This kind of sucks because I have two conventions for doing set based operations and I have to pass the data context to the static class. I've seen other people recommending piling helper methods into a partial of the datacontext like; myDataContext.ClearTemporarySelections(); But I feel this makes the dc a dumping ground for in-cohesive operations. Surely I'm missing something. I hope so. What's the convention? TIA

    Read the article

  • How do I put array into columns

    - by mathew
    $db = mysql_connect("", "", "") or die("Could not connect."); mysql_select_db("",$db)or die(mysql_error()); $sql = "SELECT * FROM table where 1"; $pager = new pager($sql,'page',6); while($row = mysql_fetch_array($pager->result)) { echo $row['persons']."<br>"; } mysql_close($db); above code output : Mathew Thomas John Stewart Watson Kelvin What I need is it should split inot multiple columns say: Mathew Stewart Thomas Watson John Kelvin HOw do I do this??

    Read the article

< Previous Page | 87 88 89 90 91 92 93 94 95 96 97 98  | Next Page >