Search Results

Search found 13451 results on 539 pages for 'physical environment'.

Page 107/539 | < Previous Page | 103 104 105 106 107 108 109 110 111 112 113 114  | Next Page >

  • Cassandra Production ready on Windows?

    - by BlackTea
    Question anyone know of any success stories of Cassandra running on windows in a production environment? I'm doing some work on Cassandra and trying to find the correct platform for it currently the platform is windows running MS-SQLas the data store. what are the dis-advantages if any when running Cassandra on a windows environment.

    Read the article

  • SQL Clustering on Hyper V - is a cluster within a cluster a benefit.

    - by Chris W
    This is a re-hash of a question I asked a while back - after a consultant has come in firing ideas in to other teams in the department the whole issue has been raised again hence I'm looking for more detailed answers. We're intending to set-up a multi-instance SQL Cluster across a number of physical blades which will run a variety of different systems across each SQL instance. In general use there will be one virtual SQL instance running on each VM host. Again, in general operation each VM host will run on a dedicated underlying blade. The set-up should give us lots of flexibility for maintenance of any individual VM or underlying blade with all the SQL instances able to fail over as required. My original plan had been to do the following: Install 2008 R2 on each blade Add Hyper V to each blade Install a 2008 R2 VM to each blade Within the VMs - create a failover cluster and then install SQL Server clustering. The consultant has suggested that we instead do the following: Install 2008 R2 on each blade Add Hyper V to each blade Install a 2008 R2 VM to each blade Create a cluster on the HOST machines which will host all the VMs. Within the VMs - create a failover cluster and then install SQL Server clustering. The big difference is the addition of step 4 whereby we cluster all of the guest VMs as well. The argument is that it improves maintenance further since we have no ties at all between the SQL cluster and physical hardware. We can in theory live migrate the guest VMs around the hosts without affecting the SQL cluster at all so we for routine maintenance physical blades we move the SQL cluster around without interruption and without needing to failover. It sounds like a nice idea but I've not come across anything on the internet where people say they've done this and it works OK. Can I actually do the live migrations of the guests without the SQL Cluster hosted within them getting upset? Does anyone have any experience of this set up, good or bad? Are there some pros and cons that I've not considered? I appreciate that mirroring is also a valuable option to consider - in this case we're favouring clustering since it will do the whole of each instance and we have a good number of databases. Some DBs are for lumbering 3rd party systems that may not even work kindly with mirroring (and my understanding of clustering is that fail overs are completely transparent to the clients). Thanks.

    Read the article

  • Looking for a target that works like "_CopyWebApplication" but for console apps

    - by Rihan Meij
    Hi We all ready have build scripts that creates our web application folders very nicely. We create multiple folders for each environment, and then change the configs in those folders according to the environment. How can we get the same results as what _CopyWebApplication does? Example: <MSBuild Projects="$(SourceCodeCheckoutFolder)\source\UI\$(ProjectName)\$(ProjectName).csproj" Targets="ResolveReferences; ResolveProjectReferences; _CopyWebApplication" ToolsVersion="3.5" StopOnFirstFailure="False" RunEachTargetSeparately="False" </MSBuild

    Read the article

  • Error 424 from Twitter API

    - by Ted
    I'm experitmenting with the Twitter API (OAuth). I have a test bed that can tweet from my dev box, but fails in production. I get a status code of 424 returned when trying to tweet. I can't spot the difference between my live environment and my dev environment so really need to understand what a '424' is, but can find no documentation relating to it.

    Read the article

  • VMWARE V2V migration without shared storage

    - by TheCleaner
    I would like to do the following: Take Physical box running w2k8r2 and Sql2008R2 and do a P2V on it to a 4.1 Enterprise licensed cluster. --no worries here, I can do that part-- Take the existing physical box that is freed up and install vmware hypervisor 5.0 on it. --again I can do this part-- Do a v2v migration of the VM created in step #1 above from the Enterprise 4.1 Cluster to the standalone host. They are NOT using shared storage. Step #3 is where I'm confused as to what my best option is. I found an article online talking about using Veeam FastSCP and just shutting down the vm on the cluster, removing it from inventory, copying the files over to the new host and then adding it to inventory. Is that the best way to accomplish this?

    Read the article

  • Moving default web site to another drive

    - by Chadworthington
    I set the default location from c:\inetpub\wwwroot to d:\inetpub\wwwroot but when I access my .NET 4.0 site get this error: Description: An error occurred during the processing of a configuration file required to service this request. Please review the specific error details below and modify your configuration file appropriately. Parser Error Message: Unrecognized attribute 'targetFramework'. Note that attribute names are case-sensitive. Source Error: Line 105: Set explicit="true" to force declaration of all variables. Line 106: --> Line 107: <compilation debug="true" strict="true" explicit="true" targetFramework="4.0"> Line 108: <assemblies> Line 109: <add assembly="System.Web.Extensions.Design, Version=4.0.0.0, Culture=neutral, PublicKeyToken=31BF3856AD364E35"/> When I try to Manage the Basic Settings on the Site and click the "Test Settings" button, I see that I have a problem under "authorization:" The server is configured to use pass-through authentication with a built-in account to access the specified physical path. However, IIS Manager cannot verify whether the built-in account has access. Make sure that the application pool identity has Read access to the physical path. If this server is joined to a domain, and the application pool identity is NetworkService or LocalSystem, verify that <domain>\<computer_name>$ has Read access to the physical path. Then test these settings again. 1) Do I need to grant rights to IIS to the new folder? Which user? I thought it was something like IIS_USER or something similar but I cannot determine the correct name of the user. 2) Also, do I need to set the default version of the framework somewhere at the Default Site level or at the Virtual folder level? How is this done in IIS6, I am used to IIS5 or whatever came with XP Pro. 3) My original site had a subfolder under wwwroot called "aspnet_client." How was this cleated? I manually copied it to the corresponding new location. My app was using seperate ASP specific databases for storing session state and role info, if that is relevant. Thanks

    Read the article

  • New iPhone Dev policy...how does Apple enforce this? [closed]

    - by milesmeow
    Apple doesn't want anyone to create iPhone apps outside of the Xcode/Objective-C environment. How can they actually enforce this? If the non Xcode IDE, for example Unity, compiles to an iPhone executable, how will Apple know which dev environment you used to create the app? Can they have Xcode compile some sort of signature into the executable that no one knows about?

    Read the article

  • Double request from mod-rewrite

    - by Dave
    I've written a module that sets Apache environment variables to be used by mod-rewrite. It hooks into ap_hook_post_read_request() and that works fine, but if mod-rewrite matches a RewriteRule then it makes a second call to my request handler with the rewritten URL. This looks like a new request to me in that the environment variables are no longer set and therefore I have to execute my (expensive) code twice for each hit. What am I doing wrong, or is there a work around for this? Thanks

    Read the article

  • Determine if string is newline in C#

    - by paradox
    I am somehow unable to determine whether a string is newline or not. The string which I use is read from a file written by Ultraedit using DOS Terminators CR/LF. I assume this would equate to "\r\n" or Environment.NewLine in C#. However , when I perform a comparison like this it always seem to return false : if(str==Environment.NewLine) Anyone with a clue on what's going on here?

    Read the article

  • What is Problem in Runtime.getruntime which does not open notepad.exe

    - by magh
    when try to execute the servlet containing following code Runtime rt = Runtime.getRuntime(); Process p = rt.exec("notepad.exe"); It doesn't launch the application in the environment windows server 2003 and Tomcat 5 but a process is being created as notepad.exe in the windows task manager. In windows xp and tomcat 6 environment launches the notepad i need to work on the windows server 2003 and tomcat 5 Thanks in advance

    Read the article

  • How to debug high memory usage by registry?

    - by bkr
    I have a windows 2008 r2 server running ADFS2 and some web apps that is having issues with low memory. Digging around I found that the 5.5 GB were being used under Kernel Memory (paged). Further digging with Poolmon, I discovered that the majority of that (5GB+) was being used by CM - configuration manager. Also known as the registry. I'm really now sure how to tell why the registry is using so much memory however, or how to release it? Looking at the physical registry files they don't appear that large. EDIT #1 Using the powershell script @ http://jdhitsolutions.com/blog/2011/05/get-registry-size-and-age/ confirms what I saw looking at the physical registry files, that they're relatively small Computername : (removed) Status : OK CurrentSize : 67 MaximumSize : 2048 FreeSize : 1981 PercentFree : 96.728515625 Created : 4/1/2011 11:38:02 AM Age : 454.23:41:28.2540682

    Read the article

  • A space-efficient guest filesystem for grow-as-needed virtual disks ?

    - by Steve Schnepp
    A common practice is to use non-preallocated virtual disks. Since they only grow as needed, it makes them perfect for fast backup, overallocation and creation speed. Since file systems are usually based on physical disks they have the tendency to use the whole area available1 in order to increase the speed2 or reliability3. I'm searching a filesystem that does the exact opposite : try to touch the minimum blocks need by an aggressive block reuse. I would happily trade some performance for space usage. There is already a similar question, but it is rather general. I have very specific goal : space-efficiency. 1. Like page caching uses all the free physical memory 2. Canonical example : online defragmentation 3. Canonical example : snapshotting

    Read the article

  • Access VirtualBox-ed server from behind the router

    - by migajek
    I'm having the following configuration: Windows 7 hosting VirtualBox and it's guest: Ubuntu The machine (physical) which runs VirtualBox is behind the router and has the address of 192.168.0.110 VirtualBox runs the Bridged network, and the IP of VirtualBox-ed Ubuntu (eth0) is 192.168.0.200 Host (Win7) is running HTTP service on port 80, while guest (Ubuntu) is running it's service on port 9000 I can access both services from inside the network by typing the ip_address:port and this works fine. Both ports are forwarded on the router to their's respective IPs: 80 -> 192.168.0.110:80 9000 -> 192.168.0.200:9000 Unfortunately, accessing the router's external IP doesn't work as expected. While external_ip:80 works correctly, external_ip:9000 - doesn't I believe the problem is VBox - related, since the same network is running also other physical machine with Ubuntu and http service on 8000 and this one is forwarded correctly.

    Read the article

  • How do I turn off automatic saving of a vim file with a ~ suffix

    - by Leonard
    In my environment, I share vim configuration with other developers (and have my own configuration additions as well) in various .vimrc files. Some places in the environment, I edit a file in vim and automagically a copy of that file with a trailing tilde suffix appears. What vim options control this? I'd like to turn the feature off, as it just clutters up my directories and spoils auto-completion on the command line. Thanks.

    Read the article

  • Help me understand how Xen vCPU/CPU works

    - by luckytaxi
    Say I have a Dual-Core server, that's 4 cores w/ two physical processors. I read numerous articles that states the dom0 should get one physical core to itself. By core, does that mean a single CPU core or one of the 4 logical cores? Ideally I would like to dedicate a single CPU core (2 logical) to the dom0. Then I would give the other CPU split between the 3 VMs. I've seen examples where ppl would assign more than the available number of cores to a VM and I don't know what good that would do. I mean, why would I want to assign 4 vCPU to a single VM when I only have 2 available (if my math is correct)? I assume I only have 2 available from the one core as I've given dom0 a CPU to itself.

    Read the article

  • How can I VPN into a domain and act as if it is local?

    - by Duall
    I've looked and looked and the internet seems to have finally failed me, there's no information on my specific issue. I have two locations, and one server (sbs 2008); I need them to act like one physical LAN. I know it's possible, I just know how to set it up, or anything about VPNs, etc. So, just to clarify a bit more: 1 Small Business Server 2008 ~30 Computers with varying systems from Windows XP to 7 2 Physical Locations I use redirected folders, but since that would consume a lot of bandwith over VPN, I have a machine over there that can be the workhorse for storage. Users need to access e-mail and shared databases from this server (or possibly some form of DFS with aforementioned workhorse) from the other building. Is there a nice tutorial (or I nice person that wants to write it) online that can explain how to set up computers from an external network to act like they were in the LAN?

    Read the article

  • Error while trying to configure VLAN with Open vSwitch

    - by Abhishek Chanda
    I am trying to connect a number of VMs using Open vSwitch. The VMs are on two separate physical boxes. Here is what I did: Created two tap devices on one physical host sudo ip tuntap add mode tap tap0 and same for tap1 Bring them up sudo ip link set tap0 up Add them as ports to Open vSwitch (br0 was created as a OVS bridge previously) sudo ovs-vsctl add-port br0 tap0 tag=1 Now I booted the VMs (using VirtualBox) and assigned addresses 192.168.122.11 and 192.168.122.12. Now when I try to ping one VM from another, I get the error Connect: network is unreachable. Since both VMs are connected to the switch on the same VLAN, I expect to be able to ping one from another. What is going wrong here?

    Read the article

  • VMware + SQL Server - sqlserver.exe not using both CPU cores

    - by fistameeny
    Hi, I am working on a virtual machine that runs SQL Server Express (as part of Sage Line 50 Manufacturing). The details are as follows: Physical Server (host machine) - Intel Xeon Quad Core 2.1GHz - 4GB RAM - VMDK image stored on RAID-5 500GB SATA drives (7200RPM) - Running Ubuntu 10.04 Server 64 bit - VMware Server 2 Virtual Machine - Windows Small Business Server 2003 - Allocated 2 vCPU's and 2GB RAM - Using 100GB pre-allocated flat VMDK file The problem I have is that there is process that runs in SQL Server that is CPU intensive. On the old physical server that we migrated to the virtual machine from, this would utilise both CPU cores so the sqlserver.exe process would be running 100% on each of the CPU cores. On the virtual machine, it only seems to use one of the two CPU cores, meaning that the process is much slower to run. Question Is there a way to force SQL Server (sqlserver.exe process) to use both of the CPU cores, and distribute it's load between them? Is this a VMware setting that needs changing to allow processes to use both cores?

    Read the article

  • Terminal Server 2003 Performance Troubleshooting

    - by MikeM
    Let me get your thoughts on terminal server performance problems. The server hosts average 25 users which, after running some numbers, on average use 600MB memory with their main applications running (web browser, adobe reader, IP phone client). All users are on the same LAN as server. We constantly experience slow response and short session lockups. Combined CPU usage is on average 10%. What appears strange to me is that the system shows 29GB physical memory with 25GB of it free. The page file usage is about 50% averaging 9GB used. Some server specs OS: Server 2003 32bit Enterprise with /PAE flag RAM: 32GB CPU: 2xQuad Core @ 2.27Ghz HD: RAID5 1.2GB After doing basic troubleshooting using performance monitor it leads me to believe that the performance problems are caused by the 32bit OS limitation in addressing full 32GB of physical memory even though the /PAE flag is used. Can anyone suggest something, troubleshooting steps that can lead to a more conclusive answer? Thanks

    Read the article

  • How can I start an X11 session on my headless Fedora 13 server?

    - by DR
    I have a small home server using Fedora 13 as operating system. Since the upgrade to Fedora 13 I cannot start the X11 server. (I need it to use vnc) When I try to start the server, both the nouveau and the original NVIDIA driver claim that there's no physical monitor attached (which is true) and that the X server cannot start because the initial display modes cannot be probed. I tried to manually add the display modes to xorg.conf but they seem to be ignored. Some forums suggest to simply use the VESA driver in this situation, but since I can't get it run either (different, more obscure message) I want to get it working with the nouveau driver as a matter of principle. Temporarily attaching a monitor would mean a lot of work for me ( 1 hour and currently it's almost 35°C/95°F in my home) I want to try that only if that definitely fixes the problem and stays that way if I remove the monitor again) How can I make the driver work without having a physical monitor attached? Thank you for your time and your help!

    Read the article

  • Windows RRAS multi-network routing

    - by Brent Pabst
    I am looking for advice, comments and suggestions from anyone who has used Windows RRAS (2008 R2 Pref.) as the primary routers for our multiple offices. We have multiple physical office locations and are looking into utilizing Windows Server 2008 R2 Core as redundant Active/Active routers/gateways for our network as opposed to a physical router from Cisco or Juniper, it costs a lot less! Any problems, issues or documentation anyone would recommend? We will still most likely have a firewall on the edge but the majority of our traffic will be inter-office with some external services. We will be using multiple Class B networks across our two offices.

    Read the article

  • explicitly refer to the a class without a namespace in C#

    - by JoelFan
    The code I'm working with has a class called Environment that is not in any namespace. Unfortunately if I am in a class that imports the System namespace, there is no way to refer to the custom class called Environment. I know this was an unfortunate choice and should be refactored, but is there any way I can explicitly refer to the conflicting class? In C++ it seems the way to do this is by using ::, and in Java there is something called global:: How do I do it in C#?

    Read the article

  • testing helpers with 'haml_tag'

    - by crankharder
    module FooHelper def foo haml_tag(:div) do haml_content("bar") end end end When I test this I get: NoMethodError: undefined method `haml_tag' This code is perfectly valid and works in a development/production environment. It's something to do with having the haml helpers properly loaded in the test environment. Thanks!

    Read the article

  • Nagios remote monitoring: NRPE Vs. SSH

    - by sam
    We use Nagios to monitor quite a few (~130) servers. We monitor CPU, Disk, RAM and a few other things on each server. I've always used SSH to run the remote commands, purely because it requires little to no additional config on the remote server, just install nagios-plugins, create the nagios user and add the SSH key, all of which I've automated into a shell script. I've never actually considered the performance implications of using SSH over NRPE. I'm not too bothered about the load hit on the Nagios server (It's probably over-speced for what it does, it's never been over 10% CPU), but we run each remote check every 30 seconds and each server has 5 different checks performed. I assume SSH requires more resources for each check but is there a huge difference? (I.E. enough of a difference to warrant the switch to NRPE). If it's any help, we monitor a mix of physical servers (Normally with 8, 12 or 16 physical cores) and Amazon EC2 medium/large instances.

    Read the article

< Previous Page | 103 104 105 106 107 108 109 110 111 112 113 114  | Next Page >