Search Results

Search found 28024 results on 1121 pages for 'sql 2014'.

Page 912/1121 | < Previous Page | 908 909 910 911 912 913 914 915 916 917 918 919  | Next Page >

  • MS Licensing - 3 windows machines, 30 users, how many CALs required?

    - by alex
    I'm in the middle of upgrading, and purchasing licensing for 3 of our Servers. One will be a Windows Server 2008 machine, running SQL Server 2008. The other two machines will be domain controllers, both running Windows 2003. Our organisation has 30 Users. I understand (through our reseller) that a Windows 2008 licence gives "downgrade" rights to use 2003. Realistically, for the above setup of 3 machines, will I just need one set of 30 CALs for 2008?

    Read the article

  • What laptop for dev with VS

    - by Gareth
    Im looking to spend upto £500 on a laptop in the next few days, and am now after a bit of adivce. This laptop will be used to develop personal projects on using VS2010, and SQL sever express. I know this sort of question has been asked plenty before however Im looking for specifics, has anyone bought a laptop recently to run VS on and if so what did you get? has anyone seen any good deals on laptops that fit in my price range? Any suggestions apprieciated Thanks

    Read the article

  • SSRS reports print button missing in IE 8

    - by Will Sams
    Seems like this is broken again (http://serverfault.com/questions/177268/ssrs-report-unable-to-load-client-print-control and SQL 2008 SP2 RsClientPrint ActiveX - "Unable to load client print control"). We updated all of our Windows 7 clients last Friday and it looks like the mighty "Unable to load client print control" has returned since then. Anyone aware of possible problems with the latest batch of Windows Updates?

    Read the article

  • Upload database backup from mysql to Amazon S3 or Glacier without creating local file

    - by Rubem Azenha
    Is there a tool that makes possible to backup a Mysql database to Amazon S3 or Amazon Glacier without having o create a local file with the database contents? Something like that: mysqldump -u root -ppass -h host --all-databases | magical-s3-tool s3-bucket backup-yyyy-mm-dd.sql This magical tool would use the pipe data and transfer the backup data directly to S3, without creating a local file.

    Read the article

  • How does Amazon EC2 operating system licensing work?

    - by JP
    Hello, I'm new to the cloud and EC2 and am wondering how licensing of operating systems works. Specifically, i spent some time looking at amazon machine images (ami's) and some contain windows and sql server. My question is how does the licensing work: a) Do I install my own licenses once the boxes are available? b) Are these licenses covered by my Amazon bill c) what are any costs associated with this software? Thank in advance, JP

    Read the article

  • How Do I Automatically Update My Database Nightly

    - by Russ
    Currently, every day before I start work, I complete the following procedure: ssh to the production server gzip our daily database dump file scp the gzipped dump file over to my computer gunzip the dump file dropdb mydatabase createdb mydatabase psql mydatabase < dump.sql Is it possible (I'm sure it is) to automate this process on Mac OSX? This way it is done by the time I get to work in the morning. If so, what is the quickest and easiest way

    Read the article

  • How Do I Automatically Update My Database Nightly

    - by Russ
    Currently, every day before I start work, I complete the following procedure: ssh to the production server gzip our daily database dump file scp the gzipped dump file over to my computer gunzip the dump file dropdb mydatabase createdb mydatabase psql mydatabase < dump.sql Is it possible (I'm sure it is) to automate this process on Mac OSX? This way it is done by the time I get to work in the morning. If so, what is the quickest and easiest way?

    Read the article

  • importing a mysqldump file does not import triggers due to some permission problems

    - by user51792
    Hello, Whenever I try to import a mysqldump export on another server, the triggers are never created, and if I remember correctly, I get an error messages that super user permission is required. IF I remove the definer it usually works but if there is a way I would prefer not having to edit the sql file. When I just simply copy over the mdi, mdy and frm files everything works perfectly. How could I import a mysqldump file so that triggers are created as well?

    Read the article

  • MySQL “filegroup”??

    - by Xaitec
    Coming for using Sql Server where there are file-groups, i was wondering if there is (i'm sure there is) something similar in MySQL. After all the database cant be limited to just one hard drive( if using windows that is). I've tried to search but its hard to find the something that you don't know the name of!.

    Read the article

  • Navicat 8/9 crashes my Debian/Linux

    - by meder
    I run Debian 5 + Gnome and it seems that after I made certain updates with aptitude, whenever I run a query in Navicat ( sql program that runs on WINE ), as the results are being presented linux goes into the command line and the GUI dies and it restarts and it asks me to login again ( all my programs crash ). Has anyone experienced this before? Or does anyone have a clue as to how I could go about debugging this? I suspect it's some issue with Gnome and WINE, but I'm not sure.

    Read the article

  • SharePoint Search with NLB

    - by Nick
    SharePoint MOSS 2007 on 64 bit OS and SQL. Added a new Web Front End to our farm, all sites seem to work fine - but now we've noticed that the search service has completely stopped working. It works if I change my host file to point to the original WFE, but if I use the NLB IP or the IP of the new WFE, it says "Unable to Connect to the Search Service

    Read the article

  • Can rdiff do incremental backups

    - by Mirage
    I am new to ubuntu , i have installed rdiff-backup. I have folder called sqlfiles on remote ftp server.The sql filesa are stored for last three days and then deleted. But i want to download the all copies to local computers I want to have incremental backups on my local server so that 1)If file is same then it should not be copied 2)if different , then overwrite it 3)If file is in local directory and not in FTP , then leave as it is How can i apply those rules to r-diff

    Read the article

  • Results copied from SSMS splitting across columns when pasted into Excel

    - by adolf garlic
    SSMS 2008 - Excel 2003 This isn't happening all the time, but sometimes when I 'copy with headers' from sql server management studio, the results in Excel look as though I've used the 'text to columns' wizard with 'spaces' and 'brackets' as delimiters, meaning that the results become raggedly distributed across columns making them useless I've looked in the 'grid output' query options but can't see anything which may be causing the issue (not sure if this should be on superuser?)

    Read the article

  • Backups devices for Windows Server Backup and Symantec [closed]

    - by user137841
    What is the best way to backup windows SQL, Exchange or AD servers data to? NAS, external USB , iSCSI or perhaps some other backup solution? I will not however be considering cloud backup solutions due to bandwidth restrictions and cost. Currently I find NAS devices to give the best results but clients that do not have the budget for backup software use Windows Server Backup but then they can make only 1 backup to a NAS at a time.

    Read the article

  • Windows Server 2008 R2 Firewall - Interface specific rules

    - by Mehmet Ergut
    I'm trying to define per interface rules, much like it was in Server 2003. We will be replacing our old 2003 server with a new 2008 R2 server. The server runs IIS and SQL Server. It's a dedicated server at the hosting company. We use a OpenVPN connection from the office to access SQL server, RDesktop, FTP and other administrative services. Only http and ssh is listening on the public interface. On the old server running 2003, I was able to define global rules for http and ssh, and allow other services only on the vpn interface. I can't find a way to do the same on 2008 R2. I understand that there is the Network Location Awareness service, firewall rules are applied according to the current network location. But I don't understand the purpose of this on a server. The only close solution I found is to set the scope on the firewall rule and restrict remote ip addresses to the private subnet of the office. But the ports will still be listening on the public interface. So how can I restrict a firewall rule to the connections coming from the vpn interface ? A note on this page states that scoping a rule to an interface does not exist anymore: In earlier versions of Windows, many of these command accepted a parameter called interface. This parameter is not supported in the firewall context in Windows Vista or later versions of Windows. I can't believe that they simply decided to remove a core firewall functionality that every firewall has. There must be a way to restrict a rule to an interface. Any ideas ? I'm still unable to find an adequate solution to my problem. So for now, my workaround is this: Administrative services listen on VPN IP address Firewall rules restrict the scope to the local IP address of VPN Public services listen on all interfaces, no scope restriction on firewall rules This is not optimal, if I change the IP address of the VPN, I need to edit the firewall rules too. It won't be the case if the rules were bound to the interface.

    Read the article

  • Why is domU faster than dom0 on IO?

    - by Paco
    I have installed debian 7 on a physical machine. This is the configuration of the machine: 3 hard drives using RAID 5 Strip element size: 1M Read policy: Adaptive read ahead Write policy: Write Through /boot 200 MB ext2 / 15 GB ext3 SWAP 10GB LVM rest (~500GB) emphasized text I installed postgresql, created a big database (over 1GB). I have an SQL request that takes a lot of time to run (a SELECT statement, so it only reads data from the database). This request takes approximately 5.5 seconds to run. Then, I installed XEN, created a domU, with another debian distro. On this OS, I also installed postgresql, with the same database. The same SQL request takes only 2.5 seconds to run. I checked the kernel on both dom0 and domU. uname-a returns "Linux debian 3.2.0-4-amd64 #1 SMP Debian 3.2.41-2+deb7u2 x86_64 GNU/Linux" on both systems. I checked the kernel parameters, which are approximately the same. For those that are relevant, I changed their values to make them match on both systems using sysctl. I saw no changes (the requests still take the same amount of time). After this, I checked the file systems. I used ext3 on domU. Still no changes. I installed hdparm, and ran hdparm -Tt on both systems, on all my partitions on both systems, and I get similar results. Now, I am stuck, I don't know what is different, and what could be the cause of such a big difference. Additional Info: Debian runs on a Dell server PowerEdge 2950 postgresql: 9.1.9 (both dom0 and domU) xen-linux-system: 3.2.0 xen-hypervisor: 4.1 Thanks EDIT: As Krzysztof Ksiezyk suggested, it might be due to some file caching system. I ran the dd command to test both the read and write speed. Here is domU: root@test1:~# dd if=/dev/zero of=/root/dd count=5MB bs=1MB ^C2020+0 records in 2020+0 records out 2020000000 bytes (2.0 GB) copied, 18.8289 s, 107 MB/s root@test1:~# dd if=/root/dd of=/dev/null count=5MB bs=1MB 2020+0 records in 2020+0 records out 2020000000 bytes (2.0 GB) copied, 15.0549 s, 134 MB/s And here is dom0: root@debian:~# dd if=/dev/zero of=/root/dd count=5MB bs=1MB ^C1693+0 records in 1693+0 records out 1693000000 bytes (1.7 GB) copied, 8.87281 s, 191 MB/s root@debian:~# dd if=/root/dd of=/dev/null count=5MB bs=1MB 1693+0 records in 1693+0 records out 1693000000 bytes (1.7 GB) copied, 0.501509 s, 3.4 GB/s What can be the cause of this caching system? And how can we "fix" it? Can we apply it to dom0? EDIT 2: I switched my virtual disk type. To do so I followed this article. I did a dd if=/dev/vg0/test1-disk of=/mnt/test1-disk.img bs=16M Then in /etc/xen/test1.cfg, I changed the disk parameter to use file: instead of phy: it should have removed the file caching, but I still get the same numbers (domU being much faster for Postgres)

    Read the article

  • How to merge many text files data in databse

    - by Mirage
    i have around 100 text files. The files have questions and 3 choices. FIles are like below ab001.txt -- contains question ab001a.txt -- is the first choice ab001b.txt ---is second choice ab001c.txt --- is third choice There are thousnad files like this. now i want to insert them in sql or first may in excel like First columns questions and other three columns as answers First two characters are same for soom files , looks like it signifies osme category so around every 30 questioons have same first charaters Any ideas

    Read the article

  • School Management System

    - by BoundforPNG
    I am looking for a school management system to replace a homegrown Access db. It should be able handle the following for both a Primary and Secondary school Scheduling classes Student Enrollment Allow teacher to enter grades and comments Generate transcripts and report cards Handle attendance Handle tuition billing It should store data in a server database like SQL Server and it would be nice to have a web interface. We are open to a commercial system or an Open Source system that comes with support.

    Read the article

  • i5+SSD or i7+HDD gives better performance [closed]

    - by Cas Sakal
    Which performs better for developers on a notebook(running VS.NET, SQL Server, no gaming); (A) i5 540M + SSD (Intel) or (B) i7 720M + 7200 RPM HDD (Western Digital) In short, I want to ask whether the performance difference between i5 540M and i7 720M can be compensated via using solid state drive instead of using a hard drive? Thank you, cas

    Read the article

  • How do you update an Excel file (Data Refresh and update formulas) WITHOUT opening the file?

    - by Alex
    I have an Excel file that want to update and save automatically with out having to open it or manually interact with. Manually, I open the file up and hit data refresh which goes and does a SQL query and then hit F9 for the formulas to update and then I just close/save. (I then would mail the file out to people using a perl script or use SAS JMP to run some numbers/charts and also mail them out. Basically I need to script some things which require the XLS file to be updated.)

    Read the article

  • Sharepoint 2010 moving site collection to different database error

    - by Brandon Ulasiewicz
    I am trying to move a site collection from one content database to another content database. First I used the following PowerShell command: New-SPContentDatabase -Name New_DB -WebApplication http://portal/ I confirmed that this did in fact create the DB in the SQL Server. I then used the following command: Move-SPSite http://portal/sites/hr -DestinationDatabase New_DB This generates an error stating that the "Operation is not valid due to the current state of the object" Can anyone help point me in the right direction with this? Thanks

    Read the article

  • What are basic differences bettwen operating systems working as server? [closed]

    - by poz2k4444
    I'd like to know what are the differences bettwen differents operating systems working as server and what operating system is better using for an specific task like linux, unix, windows, solaris, novell, etcetera. It's not about a specific problem, it's just to know, for example, windows server works better with SQL Server, or Ruby with MacOS It's just I think some of us don't know what to choose when we have to use some specific technology

    Read the article

< Previous Page | 908 909 910 911 912 913 914 915 916 917 918 919  | Next Page >