Search Results

Search found 95201 results on 3809 pages for 'system data sqlite'.

Page 266/3809 | < Previous Page | 262 263 264 265 266 267 268 269 270 271 272 273  | Next Page >

  • File store: CouchDB vs SQL Server + file system

    - by Andrey
    I'm exploring different ways of storing user-uploaded files (all are MS Office documents or alikes) on our high load web site. It's currently designed to store documents as files and have a SQL database store all metadata for those files. I'm concerned about growing out of the storage server and SQL server performance when number of documents reaches hundreds of millions. I was reading a lot of good information about CouchDB including its built-in scalability and performance, but I'm not sure how storing files as attachments in CouchDB would compare to storing files on a file system in terms of performance. Anybody used CouchDB clusters for storing LARGE amounts of documents and in high load environment?

    Read the article

  • PHPForm Data Generate PDF & Send to Email?

    - by tom
    I'm a beginner in PHP I was wondering if this is easy to do or if i'd have to outsource this to a programmer - Basically when a user fills in the PHP Form and submits it I need this to generate as a PDF (with all the labels etc) which will then email/attach to MY email and NOT the user who submitted this form. I have looked at tcpdf, fpdi but i dont think any of those scripts allow me to do this specifically as from what i heard it generates a download link for the user, and that is not what i need. If anyone can help me it would be greatly appreciated. Regards Tom

    Read the article

  • way to back up sql server 2005 database to just get the data and objects

    - by Caveatrob
    I'm trying to do a nightly backup of a SQL Server 2005 database to get the leanest, smallest file that includes all the table data, stored procedures, views, and functions. I'm downloading and saving this daily in case the server isn't available and I need to just get some data out. I don't care about the logs or any other parts of the DB but the data, procs, tables, and functions. What's the easiest way to get just this information backed up?

    Read the article

  • Linux: don't use file system cache under a directory

    - by GetFree
    For a PHP website I'm monitoring, I need to see what files are being used each time the browser makes a request. I thought of using find . -type f -amin 1. With that I get all files which were read in the last minute (it's a developing server so only I am using the website). I took care of removing the noatime attribute from the mounting point. However there must be something else that's preventing the kernel from reading the actual files on disk because the access time is not being updated when I read a file. I guess it must be the file-system cache which is retrieving the files from memory. Is there a way to disable file caching under a specific directory? (public_html in my case) Also I read somewhere that there is the nobh mounting atributes which apparently disables file caching under that mounting point, but I'm not sure.

    Read the article

  • does windows incremental backup include system state backup?

    - by Kossel
    I'm managing my very small office server with windows server 2008. since I have only one server, and the user group is really small. I made the first hdd into 2 partitions. one (C:) for windows and Active directory, another (D:) for tomcat and database. I'm doing incremental back C: and D: daily to hdd2 (E:) using windows server backup. is it enough to let me do fully restore my server in case of disaster? I ask this because I have read there is also a system state backup, and I also have to do that periodically in order to get AD back? isn't it with incremental/full backup I can do full bare-metal recovery?

    Read the article

  • 14.04 default locales

    - by Seán Ó Séaghdha
    After a new install of 14.04 I have quite a few extra locales/languages. Is this normal? I installed using English (GB) I think since en_AU isn't an option. At some point apt offered a list of unused files which I removed. Now when I open Language Support it warns me that "Language support is not installed completely" and offers to reinstall this list... kde-l10n-zhcn thunderbird-locale-en-gb thunderbird-locale-es libreoffice-l10n-en-gb libreoffice-help-es libreoffice-help-en-gb wspanish thunderbird-locale-zh-cn myspell-es ibus-sunpinyin mythes-en-au kde-l10n-es libreoffice-l10n-zh-cn fonts-arphic-uming ibus-table-wubi thunderbird-locale-es-es thunderbird-locale-zh-hans libreoffice-l10n-en-za thunderbird-locale-es-ar hunspell-en-ca libreoffice-l10n-es libreoffice-help-zh-cn kde-l10n-engb fonts-arphic-ukai So are Spanish and Chinese installed by default in all installations now? Why do I need Canadian English spellchecking installed?

    Read the article

  • Managing multiple references of the same game entity in different places using IDs

    - by vargonian
    I've seen great questions on similar topics, but none that addressed this particular method: Given that I have multiple collections of game entities in my [XNA Game Studio] game, with many entities belonging to multiple lists, I'm considering ways I could keep track of whenever an entity is destroyed and remove it from the lists it belongs to. A lot of potential methods seem sloppy/convoluted, but I'm reminded of a way I've seen before in which, instead of having multiple collections of game entities, you have collections of game entity IDs instead. These IDs map to game entities via a central "database" (perhaps just a hash table). So, whenever any bit of code wants to access a game entity's members, it first checks to see if it's even in the database still. If not, it can react accordingly. Is this a sound approach? It seems that it would eliminate many of the risks/hassles of storing multiple lists, with the tradeoff being the cost of the lookup every time you want to access an object.

    Read the article

  • Should you always pass the bare minimum data needed into a function

    - by Anders Holmström
    Let's say I have a function IsAdmin that checks whether a user is an admin. Let's also say that the admin checking is done by matching user id, name and password against some sort of rule (not important). In my head there are then two possible function signatures for this: public bool IsAdmin(User user); public bool IsAdmin(int id, string name, string password); I most often go for the second type of signature, thinking that: The function signature gives the reader a lot more info The logic contained inside the function doesn't have to know about the User class It usually results in slightly less code inside the function However I sometimes question this approach, and also realize that at some point it would become unwieldy. If for example a function would map between ten different object fields into a resulting bool I would obviously send in the entire object. But apart from a stark example like that I can't see a reason to pass in the actual object. I would appreciate any arguments for either style, as well as any general observations you might offer. I program in both object oriented and functional styles, so the question should be seen as regarding any and all idioms.

    Read the article

  • Is it Considered Good SQL practice to use GUID to link multiple tables to same Id field?

    - by Mallow
    I want to link several tables to a many-to-many(m2m) table. One table would be called location and this table would always be on one side of the m2m table. But I will have a list of several tables for example: Cards Photographs Illustrations Vectors Would using GUID's between these tables to link it to a single column in another table be considered 'Good Practice'? Will Mysql let me to have it automatically cascade updates and delete? If so, would multiple cascades lead to an issues? UPDATE I've read that GUID (a hex number) Generally takes up more space in a database and slows queries down. However I could still generate 'unique' ids by just having the table initial's as part of the id so that the table card's id would be c0001, and then Illustrations be I001. Regardless of this change, the questions still stands.

    Read the article

  • "Unable to install GRUB in /dev/sda" when installing GRUB

    - by vicban3d
    I recently bought a shiny new Lenovo Yoga 2 Pro and I want to dual boot it with Ubuntu for studying purposes. Its built-in OS is Windows 8.1 and it has a 256GB SSD. I've made a separate 90GB partition just for Ubuntu and a live USB to install it. The first time everything seemed to work great, I solved the wifi issued by blacklisting ideapad_laptop, the installation went flawlessly and Ubuntu worked fine. When I got up the next morning and turned on my laptop it booted into Windows right away without ever showing the GRUB menu. So I tried to reset, and checked my partitions with the Disk Manager and everything looked fine. Since I couldn't find a solution online I went ahead and formatted the partition to try and install again. This time and every time since, the installation was aborted and I got a fatal error saying: Unable to install GRUB in /dev/sda Executing `grub-install /dev/sda` failed. This is a fatal error. Can anyone please suggest a solution to this problem? If any further information is needed I would be happy to provide it. Thanks. When installing I get the following in details: ubuntu kernel: [ 1946.372741] FAT-fs (sda2): error, fat_get_cluster: invalid cluster chain (i_pos 0). ubuntu grub-installer: error: Running 'grub-install --force failed.

    Read the article

  • Which file system to choose from when formatting 1.5TB hard drive (hdd)

    - by MaxiWheat
    I plan to buy a 1.5TB hard drive soon. I would like to know which file system to choose from when I'm gonna format it. With FAT32, there is a limitation concerning the maximum file size (4GB) that bugs me since I might save large files such as DVD images which are over 4GB. On the other hand, NTFS allows me to save larger files, but seems less compatible with other OS than Windows and is also proprietary to Microsoft. Are there other alternatives ? Can you give me your advices ?

    Read the article

  • Dropbox icon dissapears right after login

    - by garvamel
    Even though Dropbox's app indicator dissapearing from the tray area seems like a recurrent enough problem, my issue is a litte different. When I login, I can see the app panel populating, and the dropbox icon does indeed appear (config'd as startup application), but after some other icons show up (bluetooth, battery, etc.) it's gone. It's still running though. I'm guessing it's having issues with staying pinned, and I don't know how to start addressing this problem. I have tried many if not all solutions provided here in the forums for the "icon missing" questions. So far: I've whitelisted everything regarding panel Uninstalled-reinstalled (with and without rebooting in between) Overwritten current installation Purged installation from terminal Installed from Software Center and from .deb file batch deleted every "dropbox" ocurrences from terminal (files and folders) and reinstalled Ran sudo apt-get install libappindicator1, it installed, but didn't solve anything I'm on Ubuntu 12.04 LTS - 64 bits. Any insight would me much appreciated!

    Read the article

  • Q: MySQL Cluster - Data insertion in NDBCLUSTER table - error out after 5 million rows

    - by Mata
    MysqlCluster version: mysql-5.6.11 ndb-7.3.2 Insertload = 50 M dataset Datanodes = 3 LOAD DATA INFILE '/input_50m/Table_1_sorted.csv' IGNORE INTO TABLE nw_ndb FIELDS TERMINATED BY ',' LINES TERMINATED BY '\n' We recently setup a new mySQL cluster and trying to load data from a flat file. But getting error “Got temporary error 4010 'Node failure caused abort of transaction' from NDBCLUSTER" when inserting 5 million rows in a single table in MySQL Cluster. We are using "LOAD DATA INFILE" command to load the data in the table from csv file. Server (musqld, ndb nodes) has good hardware: 126 GB RAM, 32 Gb allocated to mysqld tried below settings with no effect: SET autocommit=0; SET FOREIGN_KEY_CHECKS=0; SET unique_checks=0; SET GLOBAL ndb_batch_size=8*1024*1024; SET GLOBAL ndb_cache_check_time = 1000; SET GLOBAL ndb_index_stat_cache_entries = 10000000; SET SESSION BULK_INSERT_BUFFER_SIZE=256217728; SET GLOBAL KEY_BUFFER_SIZE=256217728; Any clues?

    Read the article

  • Outlook data file cannot be accessed. [closed]

    - by Alex
    Possible Duplicate: Outlook data file cannot be accessed. Hello. I am using the beta 2010 Office Outlook. When I try receive or send, i get the following error "outlook data file cannot be accessed." Repair/reinstall office and try to use any different outlook data files is not take any effect.

    Read the article

  • Looking for a central image database and tagging system for a group of users

    - by jstarek
    I'm doing IT support for a small volunteer organisation who needs to centrally store and organize around 2500 photos. Can anyone here recommend a database or similar system which matches the following criteria: Intuitive to use for users with little computer experience Multi-user support, ideally with integration in our existing LDAP user directory Should have a web-interface Not a hosted solution like Picasa (because we have a rather slow internet connection with very slow upload) Should allow tagging of images, sorting by various criteria and storing copyright information If there are native GNOME and/or Windows clients for the tool, that would be a great benefit. Many thanks in advance!

    Read the article

  • Data base preference for network based C# windows application [on hold]

    - by Sinoop Joy
    I'm planning to develop a C# widows based application for an academy. The academy will have different instances of application running in different machines. The database should have shared access. All the application instances can do update, delete or insert. I've not done any network based application. Anybody can give any useful link to where to start with ? Which database would give max performance with all required features i said for this scenario ?

    Read the article

  • Create a Windows 7 System Repair ISO?

    - by MaDMaD Mad
    According to the Microsoft website I can run "Create a system repair disc" from the start menu to burn a recovery Disc. However, there are many situations (including not having a disk drive) where this cannot be done. My goal is to create a bootable USB drive according to this guide but their method to do so without burning a physical disc includes downloading an ISO from "somebody's" website. I don't know who "somebody" is and I feel it is reasonable to not boot something I can't verify as being untampered with. To the best of my knowledge, Microsoft does not host a recovery disc ISO for download as this would be the easiest solution. So my problem is that I'd like to circumvent Microsofts "burn to disc" option with a "burn to ISO" through any reasonable means.

    Read the article

  • Ways to remotely reboot a Linux system

    - by dualed
    I had a remote server running Debian Sarge that experienced some HDD failure and I meant to reboot it hoping that fsck could repair the errors automatically. I eventually drove out there and replaced the faulty disks... But I was wondering: What other ways are there to force a Linux system to reboot that do not require hard drive access? What I had tried: shutdown -r now Did not work, as shutdown is a program that would have to be loaded from disk, the error shown in the terminal was bash: /sbin/shutdown: Input/output error init 6 same as above telinit q same as above kill -2 1 This did not print an error, but did not work either. (However, it is possible that the Sarge init did not implement SIGINT, the sarge manpages did not mention it. So it could work in a more recent version of Debian) This guide on PCFreak.net. However, this failed at sysctl, which was not in memory either.

    Read the article

  • Silverlight 4 + RIA Services - Ready for Business: Consuming Data in the Silverlight Client

    To continue our series, lets see where the fun comes in my look at how easy that is to consume from the client.  First just to help you understand what is happening behind the covers, lets look at a code-behind solution.  In View\Home.xaml put a simple DataGrid on the form. <sdk:DataGrid Name="dataGrid1" Height="152" Width="692" /> Then add these lines of code to Home.xaml.cs   1: var context = new DishViewDomainContext(); 2: this.dataGrid1.ItemsSource...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Knife leaves stray processes on my system

    - by Leons
    I'm seeing stray knife processes on my system. I have an automated ruby script that runs bundle exec knife bootstrap against various nodes. Most of the time the knife process completes and goes away, but sometimes it stays for days. I'm noticing it days later in ps aux I think it's related to the target node being down when knife runs. The chef server timeout is high, so the action completes eventually when the node goes back up, but I think knife may give up or hang somehow during the wait. Is there something I can do about the stray knife processes? Does knife have timeout settings separate from the chef server's timeout settings?

    Read the article

  • Repair ext4 filesystem on USB drive

    - by phineas
    Yet another filesystem question. I wanted to use a USB drive that I hadn't mounted for a month or so and was surprised by the fact Ubuntu was unable to mount it. I looked it up in the disk utility and it said it discovered a device with 17 MB instead of 2 GB. The hardware looks intact, I hope for the best for repairing the ext4 filesystem. I followed the instructions from HOWTO: Repair a broken Ext4 Superblock in Ubuntu, but I wasn't successful. # fsck.ext4 -v /dev/sdb e2fsck 1.42.5 (29-Jul-2012) ext2fs_open2: Bad magic number in super-block fsck.ext4: Superblock invalid, trying backup blocks... fsck.ext4: Bad magic number in super-block while trying to open /dev/sdb The superblock could not be read or does not describe a correct ext2 filesystem. If the device is valid and it really contains an ext2 filesystem (and not swap or ufs or something else), then the superblock is corrupt, and you might try running e2fsck with an alternate superblock: e2fsck -b 8193 Filesystem blocks are invalid, however when I run the recommended solution to try the alternate superblock, I get the following output: # e2fsck -b 8193 /dev/sdb e2fsck 1.42.5 (29-Jul-2012) e2fsck: Invalid argument while trying to open /dev/sdb plus the same error message as in the last paragraph above. Any ideas how to recover the drive? Thank you very much! Edit: testdisk won't help. I'm still stunned why the tools only discover 17 MB.

    Read the article

  • Alerting when a RAID Array disk fails locally on VMWare ESX or ESXi System

    - by Tim K
    With ESX and ESXi, we recently had two systems where that the boot partition became degraded due to a failed disk. The only alert we managed to capture was the visual alert on the Dell servers. We failed to received any electronic alerts regarding the failed or degraded array. Does anyone have any experience with monitoring for these types of failures? In both cases, the servers were running in a RAID 5 SCSI configuration (5 disks on one system, 3 disks on another) which if we were running a Windows Server OS, we would have had an alert created in the Eventviewer. Where would I begin to look for this solution. Can it be configured in VCenter or vFoglight?

    Read the article

  • Recovery from hell - undeleting partition overwritten by Xubuntu 12.10 installer?

    - by DaimyoKirby
    This is turning into a nightmare - following my initial recovery of my two partitions, I went to install Xubuntu 12.10 (again). At this time I had two partitions - one of ~39 GB had Zorin OS 6 installed on it, and another of ~33 GB had nothing installed, just a few files in it that I had manually backed up (moved) there. When I got to the partitioning step, I chose "Replace Zorin OS 6 with Xubuntu 12.10", along with LVM, naturally thinking that the installer wouldn't touch the second partition, since Zorin wasn't installed on it. I was dead wrong. Upon booting my newly installed Xubuntu 12.10, I found in gparted that there were only two partitions - ~255MB, which appears to have the boot stuff in it (it's flagged boot in gparted), and another of ~74 GB. Question: Is there any way to salvage my old files on the non-Zorin ext3 partition? I'm really upset I made such a dumb move (again...), and any and all help is appreciated very, very much!

    Read the article

< Previous Page | 262 263 264 265 266 267 268 269 270 271 272 273  | Next Page >