Search Results

Search found 16094 results on 644 pages for 'it helpdesk team manager'.

Page 193/644 | < Previous Page | 189 190 191 192 193 194 195 196 197 198 199 200  | Next Page >

  • Kernel Errors in logwatch

    - by Vince Pettit
    We have a dedicated server running CentOS and Plesk. We've had the following show up on our logwatch and wondered if it is anything we should worry about? --------------------- Kernel Begin ------------------------ WARNING: Kernel Errors Present Northbridge Error, node 1K8 ECC ...: 1 Time(s) ---------------------- Kernel End ------------------------- We've contacted the support team that we rent our server from but they don't seem to want to help us out without us paying their support team a fixed charge and even then they can't guarantee they would be able to find a solution to any potential problems. Full log lines regarding Kernel error... Jun 16 19:45:25 server88-208-217-241 kernel: Northbridge Error, node 1<0>K8 ECC error. Jun 16 19:45:25 server88-208-217-241 kernel: EDAC amd64 MC1: CE ERROR_ADDRESS= 0x2a3d553e0 Jun 16 19:45:25 server88-208-217-241 kernel: EDAC MC1: CE page 0x2a3d55, offset 0x3e0, grain 0, syndrome 0x5041, row 3, channel 0, label "": amd64_edac

    Read the article

  • Does shutdown idle VMs improve the performance?

    - by Samselvaprabu
    Often our team members are coming to me with a compliant that their VMs are slow. Our team members suggested to shutdown some of the VMs temporarily and try to access the VM. But most cases that would not help. Assume that i have assigned 4 GB for and 2 CPUs for my VM. So ideally it should not face performance issue. As our ESXi 4.1 server has multiple VM in the same server (we have overcommited memory and CPU). Does shut down other VM really helps to improve performance or not? [Note : We are using ESXi 4.1 and our hardware is R710 server. We have more number of VMs in single server so we have overcommited memory.]

    Read the article

  • Linux stretch cluster: MD replication, DRBD or Veritas?

    - by PieterB
    For the moment there's a lot of choices for setting up a Linux cluster. For cluster manager: you can use Red Hat Cluster manager, Pacemaker or Veritas Cluster Server. The first one has the most momentum, the second one comes by default with RH subscriptions and the last one is very expensive and has a very good reputation ;-) For storage: - You can replicate LUN's using software raid / md device - You can use the network using DRBD replication, which offers a bit more flexibility - You can use Veritas Storage Foundation technology to talk to your SANs replication technology. Anyone has any recommandations or experience with these technologies?

    Read the article

  • Set up DPM to use hardware VSS Provider with EMC Clariion SAN

    - by Ryan
    We recently purchased an EMC Clariion AX4 SAN and we're using it to provide storage for our Hyper-V virtual machines. I've already got the servers registered with it and configured the failover cluster CSV's, etc. I'm wanting to figure out what I have to do to set up the VSS Provider for the SAN, so that Microsoft Data Protection Manager can use it to backup the virtual machines. I'm new when it comes to Clariion SAN's, so I have no idea how to set up the VSS provider. I downloaded something from Powerlink that was labeled VSS Provider and installed it on one of the machines connected to the SAN. It had something to do with a Solutions Enabler (I don't know what that is), but it doesn't really seem like it did anything. I read something that suggested I'd need to have Navisphere Manager to use the VSS Provider on the AX4, but we didn't purchase that - we're just using Navisphere Express. Can anyone help me figure out how to get the VSS Provider up and running?

    Read the article

  • Should app config files be in jboss/bin directory?

    - by djangofan
    I am a Integration Engineer for a software company. My development team has generated various jboss .ear applications that rely on configuration files being in the jboss/bin directory and setting the CLASSPATH externally using batch files (on Windows from an external LIB directory) rather than using jboss's internally loaded classes. Is this the right/standard way of doing things? Would some other location be better (for the config files) such as jboss/server/instancename/deploy or some other directory? Any opinons on how I can direct my development team to doing things in a best practices fashion?

    Read the article

  • Connecting to local Sql server 2005 through Internet

    - by Manish
    Hello My Sql server is on Local Machine, I want to access it through Internet. I Configure Surace manager and Configuration manager of sqlserver 2005 for remote access. My sqlserver is running on port:1433 I am using port forwarding , I can access service of other ports, My Connetion String Is: Data Source=190.190.200.100,1433;Network Library=DBMSSOCN;Initial Catalog=myDataBase;User ID=myUsername;Password=myPassword; But it gives following error when i am trying to connect sql server through internet: A network-related or instance-specific error occurred while establishing a connection to SQL Server. The server was not found or was not accessible. Verify that the instance name is correct and that SQL Server is configured to allow remote connections. (provider: TCP Provider, error: 0 - A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond.)

    Read the article

  • Weird Windows 2003 MSDTC and SQL 2005 issue

    - by seagull surfer
    scenario: Windows 2003 sp2 x64 enterprise edition. SQL 2005 sp2 cu9 x64 Enterprise edition After restarting the resource groups on two node active-active cluster, 3 SQL 2005 instances start up fine. The 4th one starts up but starts throwing the following error. "Enlist operation failed: 0x8004d00e(XACT E NOTRANSACTION). SQL Server could not register with Microsoft Distributed Transaction Coordinator (MS DTC) as a resource manager for this transaction. The transaction may have been stopped by the client or the resource manager." MSDTC is fine since the other 3 function normally. The only way to "fix" it is to take the 4th instance offline and bring it online again. Is there any way to fix this enlistment without restarting?

    Read the article

  • How to get Windows Server 2008 VM to use multiple cores

    - by David Fraser
    I have a Windows Server 2008 machine running in VirtualBox. On initial installation, only one processor was made available, but now I want to run it as a multiprocessor machine. I have made all four cores available in the VirtualBox settings (as well as enabling VT-x/AMD-V and Nested Paging), but Task Manager still only shows one CPU. However, the four CPU cores are visible in Device Manager under Processors. In the event log on startup, I can see the following relevant events: EventLog.6009 Microsoft (R) Windows (R) 6.00.6002 Service Pack 2 Multiprocessor Free Kernel-Processor-Power.4 Processor 0 exposes the following: 1 idle state(s), 0 performance state(s), 0 throttle state(s) Kernel-Processor-Power.4 Processor 255 exposes the following: 0 idle state(s), 0 performance state(s), 0 throttle state(s) Kernel-Processor-Power.4 Processor 255 exposes the following: 0 idle state(s), 0 performance state(s), 0 throttle state(s) Kernel-Processor-Power.4 Processor 255 exposes the following: 0 idle state(s), 0 performance state(s), 0 throttle state(s) How can I make this system actually boot up as a multiprocessor machine?

    Read the article

  • Multiple test Active Directory envirovments hand in hand with production domain controllers

    - by MadBoy
    What's the best approach of having multiple test environments next to production one? We have multiple programming teams that build solutions that use Active Directory very often. We have tried different approaches, starting with their own domain controllers (in same subnet), or additional OU's in our production AD that the team gets control over and can create/delete accounts within that one OU. We thought of possible 4 solutions: Setting up separate OU's in ou production env. Creating subdomains for our contoso.com domain like test.contoso.com, something.contoso.com and delegating control to the teams (would we need additional DC's or the two that we have already would be enough to hold this? Setting up additional test domain controler that has a trust to our main domain and all teams can use the test domain controler as they please. Setting up single domain controller for every team/project. We're taking in consideration amount of resources needed, security (for example having multiple domain controlers with multiple passwords may lead users to use simpler passwords) and overall best practices for this scenario.

    Read the article

  • What's the best way to do user profile/folder redirect/home directory archiving?

    - by tpederson
    My company is in dire need of a redesign around how we handle user account administration. I've been tasked with automating the process. The end goal is to have the whole works triggered by the business, and IT only looking in when there's an error reported. The interim phase is going to be semi-manual. That is a level 2 tech inputs the user's info and supervises the process. The current hurdle I'm facing is user profile archiving. Our security team requires us to archive the profile directories for any terminated user for 60 days in case the legal team requires access to their files. Our AD is as much a mess as everything else, so there are some users with home directories and some with profiles. Anyone who has a profile dir in AD also has a good deal of their profile redirected to our file servers over DFS. In order to complete the process manually you find the user in AD, disable them, find their home/profile dir, go there and take ownership, create an archive folder, move all their files over, then delete the old dir. Some users have many many gigs of nonsense and this can take quite some time. Even automated the process would not be a quick one. I'm thinking that I need to have a client side C# GUI for the quick stuff and some server side batch script or console app to offload this long running process. I have a batch script that works decently using takeown and robocopy, but I wonder if a C# console app would do a better job. So, my question at long last is, what do you think is the best way to handle this? I can't imagine this is a unique problem, how do other admins get this done? The last place I worked was easily 10x larger than the place I'm in now. If we would have been doing this manual crap there, they'd have needed a team of at least 30 full time workers to keep up. I have decent skills in C#.net and batch scripting, but am a quick study and I have used most every language once or twice. Thank you for reading this and I look forward to seeing what imaginative solutions you all can come up with.

    Read the article

  • Problem with USB drivers (Windows-XP)

    - by Carl
    I obtained the drivers from the manufacturer for my HT-Link NEC USB 2.0 2-port Cardbus card. When I plugged in the card before I got the drivers, 3 new entries showed up in the Device Manager - two "NEC PCI to USB Open Host Controller" and one "Standard Enhanced PCI to USB Host controller." With the card plugged in, I uninstalled those two drivers. I then removed the card. I copied the new drivers to c:\windows\system32\drivers and the .inf file to c:\windows\inf. I also copied the drivers & inf to a new directory called c:\windows\drivers\ousb2. I reinserted the card. Windows automatically installed the same drivers as before. I selected 'update driver' on the "NEC PCI to USB..." entry and didn't see any other options. I then selected 'have disk' and pointed to c:\windows\drivers\ousb2 and got a message "The specified location does not contain information about your hardware." I then selected 'update driver' on the "Standard Enhanced PCI to USB...," and manually selected "USB 2.0 Enhanced Host Controller" (OWC 4/15/2003 2.1.3.1). Windows then automatically found a USB root hub, and I manually selected "USB 2.0 Root Hub Device" (OWC 4/15/2003 2.1.3.1). Now there are two sections in the Device Manager titled "Universal Serial Bus controllers." I plugged in my external USB hard disk adapter, and "USB Mass Storage Device" was added to the first set. Here's how it looks (w/drivers from the properties): [Universal Serial Bus controllers] Intel(R) 82801DB/DBM USB 2.0 Enhanced Host Controller - 24CD (6/1/2002 5.1.2600.0) Intel(R) 82801DB/DBM USB Universal Host Controller - 24C2 (7/1/2001 5.1.2600.5512) Intel(R) 82801DB/DBM USB Universal Host Controller - 24C4 (7/1/2001 5.1.2600.5512) Intel(R) 82801DB/DBM USB Universal Host Controller - 24C7 (7/1/2001 5.1.2600.5512) NEC PCI to USB Open Host Controller (7/1/2001 5.1.2600.5512) NEC PCI to USB Open Host Controller (7/1/2001 5.1.2600.5512) USB Mass Storage Device USB Root Hub (7/1/2001 5.1.2600.5512) (5 more USB Root Hubs - same driver) [Universal Serial Bus controllers] USB 2.0 Enhanced Host Controller (OWC 4/15/2003 2.1.3.1) USB 2.0 Root Hub Device (OWC 4/15/2003 2.1.3.1) When I unplug the card the two "NEC PCI to USB..." entries in the first set disappear, and the whole second set disappears. (I unplugged the hard disk adapter first...) The hard disk adapter still doesn't work in that Cardbus card with the new drivers. I don't think the above looks right - a second set of USB controllers listed in the Device Manager, and the NEC entries still in the first set, and the the USB mass storage device still in the first set. Any help appreciated. (Windows XP PRO SP3 w/all current updates.)

    Read the article

  • non-GUI connection to local Hyper-V VM without network

    - by sandro
    I have a virtual machine on Hyper-V manager (Windows 2008 R2) without a network configured on the VM. From a powershell script running on the host Windows server, I would like to query into the OS of that local VM for certain information (i.e. if a given process has finished completion). I am using codeplex's pshyperv module (https://pshyperv.codeplex.com/) to interact with Hyper-V manager, but the only cmdlet to connect to the vm is 'New-VMConnectSession', which launches a 'vmconnect.exe' connection to the VM. Since vmconnect.exe is essentially RDP, this is not very script-friendly. From within a host's powershell script, is there any way to send a command to a local virtual machine's OS and receive output, if no network is configured on the VM? (I believe Vmware's 'vmrun' utility has this capability) Another way to ask this question: Does Hyper-V have a non-GUI-based form of vmconnect.exe? (PS. Not sure if this was more stackoverflow or serverfault)

    Read the article

  • File permissions issue with an NFSv4 share, uploaded from a Mac Lion

    - by POP.sicle
    I have an NFSv4 share that was working fine, with Macs using Snow Leopard, to share files across the network. The NFS share has one cloned user/group that all clients autoconnect as. However, when I use a Lion Mac to copy a file from their user directory to the NFS, no other computer (mac SL/mac Lion/Win7) can edit/delete/write to the file that was uploaded - despite having the correct read/write/ex permissions visible on the NFS and through terminal. Attempting to edit the file permissions through Finder completely locks the file. I suspect this has something to do with Lion's ACLs (or maybe its version control) conflicting with NFSv4. Is there a way to disable or ignore extended ACLs or extended file permissions on the NFSv4 side, that would allow users to not run into this conflict? The work around currently is to use NFS Manager and set automounts to ignore ownership but installing NFS manager and configuring automounts for all of the computers seems more troubling than attempting to reconfigure the NFS settings. Advice?

    Read the article

  • use yum behind a proxy on CentOS 5.6

    - by ewok
    I have a CentOS VM running in Virtualbox on a Windows 7 host (all 64-bit). I have configured the proxy in the System menu and have added the proxy=http://my.proxy.com:12345 line to my /etc/yum.conf file. I am able to connect to the internet from a browser, and I can update packages form the update manager and install them from the package manager. However, I cannot use yum for anything. When I attempt to use yum (install, update, upgrade, or anything else), It attempts to connect to the mirrors, does nothing, then after a minute or so, quits. Does anyone know how to configure yum to work behind a proxy?

    Read the article

  • Is there a difference between starting a WebLogic Admin Server through the command-line or through t

    - by lindelof
    I have WebLogic 11g installed on my machine under /srv/wls, and I start Node manager on it with /srv/wls/wlserver_10.3/server/bin/startNodeManager.sh. I create my domain with config.sh, and then I want to start the Admin Server of that domain with Node Manager. So I open a WLST session and do the following: > nmConnect(...) > nmStart('AdminServer') But when I then visit localhost:7001/console, and login, the screen on the browser remains blank and the server logs tell me the server has run out of PermGen space. I don't encounter this problem when starting the Admin Server through the shell script under $DOMAIN/startWebLogic.sh. Then it works fine. Any idea what I am doing wrong? I couldn't find anything in Google nor on the Oracle forums...

    Read the article

  • No HDMI audio - Windows 8 - ASUS H81M-PLUS

    - by Paul Wright
    I have an issue with HDMI audio on Windows 8 using an ASUS H81M-PLUS motherboard (without an external GFX card). There are many forum posts advising you to go into playback devices and setting HDMI to be default - I have done this. To eliminate what works and what doesn't work: I have not been able to get sound from my HDTV using HDMI. I have used this HDMI cable with my PS3, so this cable should be fine. I am able to use the HDMI cable in extended mode, so that I have two monitors (including the TV), just no audio. This HDMI cable goes straight from the motherboard to the TV. Below I have included 'Device manager', and 'Playback Devices' (Sound). Device Manager Playback Devices, showing disabled and disconnected devices I am at a loss. I have uninstalled all drivers, and then rebooted and made windows look for the correct ones, made sure the HDMI device was default. Thanks, Paul

    Read the article

  • Question about Linux Device Drivers.

    - by smwikipedia
    I have just installed the Xubuntu. My box runs quite slower than before when installed with a Win7. I am wondering if I am missing some device drivers. I can use gnome-device-manager to list my devices. In the device list view, I saw several blue question mark on the left side of devices. Based on my windows experience, the question mark should indicate the device is not properly configured. Is it the same with gnome-device-manager? And most importantly, how could I know for sure whether or not the corresponding device drivers have been installed? Many thanks.

    Read the article

  • Bluetooth radio device is not available

    - by Mike
    I reinstalled my Windows 7 operating system, and have since been unable to detect my Logitech m555b bluetooth mouse. Here is some info about my system: I have have a working connection with a bluetooth printer. I have a message in the "My Bluetooth" section of explorer stating "Bluetooth radio device is not available". The Device Manager indicates that the Generic Bluetooth radio is working correctly. The bluetooth "F12" switch is on. Mouse batteries are new and the right way around. I'm presuming that I need a bluetooth driver which is non generic. The computer is a Clevo P150HMx (the version with the GTX485 GPU) with the default WLAN / bluetooth combo card manufactured by Realtek (I think it's a Realtek RTL8188CE card). I think I have the drivers installed, but I still get the generic driver in the Device Manager. I'm confused. Please help, I'm going mad on the touchpad. (Thanks for the touchup wizlog)

    Read the article

  • Hyper-V management remotely

    - by Péter
    I'll tell you in advance that I'm newbie in the topic. I have a Win8 (Home) machine with Hyper-V installed behind a router. The router has a public IP and a domain attached. I have another Win8 (Work) machine also installed Hyper-V. I want to access to my home Hyper-V via Hyper-V Manager so I can manage my virtual machines from work. I found this article but I don't know if it's applicable to me. I thought that a simple port forwarding should work and I only need to do is grant the Work HV manager my domain and the port I choose and if it's pop a login form I only need to fill the user data of my Home computer? How can I solve this? My thoughts revolve around: - Port forwarding - set domain+port and set my home user - Set up a VPN and use the local ip address of my home computer (it looks like a little cumbersome and my router only support PPTP) I'm open to any other solution too. Thanks, Péter

    Read the article

  • Routing from the DMZ to the interior network only

    - by Allan
    I have a home network connected to Verizon's FIOS service. Verizon's ActionTec router is connected to the ONT via coax to establish the MOCA network. My DD-WRT router's WAN port is connected to one of the ActionTec's LAN ports. The DD-WRT router is configured with a static IP address and assigned to the DMZ (this is done so I only have to configure port forwarding once). My issue is that this does not allow computers connected to the DD-WRT to serve streaming audio/video to the MOCA network using Verizon's Media Manager. I know that Media Manager uses ports 18001, 5050, and 5060, but I don't know how to forward those ports so that they are only available to the ActionTec's network and not the rest of the internet.

    Read the article

  • Renting an "EC2" server VS buying one (for a start up in initial stages)

    - by krish p
    We are a small start up in the early stages and are working on a SaaS-based Rails product. Currently, we use EC2 for a small instance and have a need for another large/extra-large instance as we are beginning to deploy to the Cloud and get ready to release our "alpha" version. While EC2 was my choice for numerous reasons (reliability, accessibility - small team is geographically dispersed, maintainability, and things of that nature), it appears to be rather expensive. While the product will ultimately be deployed in the Cloud (be it EC2 or otherwise) and that experience would help the development team, would it make sense to purchase a physical server and stick it in the basement or bite the bullet and pay the price for EC2 (or other Cloud Providers)? While such decisions are driven by numerous factors, it would certainly help to get the thoughts of other folks who may have been in similar situations. Hence, the post. Thanks much!

    Read the article

  • Help me find the offending process waking my Windows 7 PC from hibernate every night

    - by DavidGugick
    Recently my Windows 7 64-bit PC has started waking every night from hibernation around 3:30am. I have done the following to try and figure out what is causing the issue with no luck: Examined the Windows Event logs. Nothing is noted Ran powercfg -lastwake and that reports nothing c:\powercfg -lastwake Wake History Count - 1 Wake History [0] Wake Source Count - 0 Ran powercfg to find what devices are armed for wake. Interestingly, this reports two items (I've already unchecked the "Allow this device to wake the computer" in device manager): The keyboard and something called the "eHome Infrared Receiver (USBCIR)". This is a desktop PC and it does not have an Infrared received, so I'm not sure what that device is. Suffice to say it does not have the option to "Allow device to wake..." available in Device Manager. C:\powercfg -devicequery wake_armed eHome Infrared Receiver (USBCIR) HID Keyboard Device My next step is to disable the Keyboard from wake, but I'm not convined that's the problem. This is on a Dell XPS435 if that helps anyone.

    Read the article

  • Adding Thunderbird-stable repository gives "can't find signing_key_fingerprint" error

    - by EBV2010
    I'm trying to install Thunderbird 11 on Kubuntu 10.04. I was able to do it on the machine I'm working on. To get a clean process that I can roll out to other clients, I re-installed the machine and repeated the process. This is what I did (I've left out the sudo for clarity): add-apt-repository ppa:ubuntu-mozilla-security/ppa apt-get update add-apt-repository ppa:mozilla-team/thunderbird-stable The last one resulted in this error: Error: can't find signing_key_fingerprint at https://launchpad.net/api/1.0/~mozilla-team/+archive/thunderbird-stable The machine as it was before re-installation gave no such message. It was built from the same sources. Bottomline: I got Thunderbird 11.0 to run on Kubuntu 10.04 but after re-installation, adding the repository gives an error and won't add. Is there a way to solve the signing_key_fingerprint error?

    Read the article

  • KDE doesn't start up anything else

    - by Shane
    I just installed KDE under Arch Linux. Problem is, nothing is starting up right with it - no window manager, no panels, nothing. All I get is a small terminal window in the bottom right corner of the screen, which I'm assuming is konsole. From that single window I can do things like start kwin or launch programs whose names I happen to know (like chromium or firefox), but I don't have a panel for starting programs or switching between programs. It doesn't matter whether I start kdm through inittab or manually by typing # /etc/rc.d/kdm start. KDM looks great, but once I log in as my normal user I just get a console window with no decoration. Is there a startup script for KDE somewhere that needs to run, and usually has a bunch of programs by default - like a window manager, panels, widgets, and all the normal background programs that run in an ordinary GUI? If so, how can I "restore" the default behavior?

    Read the article

< Previous Page | 189 190 191 192 193 194 195 196 197 198 199 200  | Next Page >