Search Results

Search found 5643 results on 226 pages for 'machines'.

Page 143/226 | < Previous Page | 139 140 141 142 143 144 145 146 147 148 149 150  | Next Page >

  • How do I force a 64-bit windows 8.1 install?

    - by ausairman
    I have just purchased a copy of Windows 8.1 32-bit/64-bit from the MSDNAA program through my University. It came as an .iso file so I burned it to a DVD and installed / registered without any issues. I just found out it's a 32-bit install though, and I'm running a 64-bit laptop (Dell Studio 1555). The exact produce name I purchased is Microsoft Windows 8.1 Professional 32/64-bit (English). Is that just another name for "32-bit version that also works on 64-bit machines"? I have 4GB and I would like to code for 64-bit architectures so not having a 64-bit environment will be a real pain. I didn't see any options to choose the install type (it had Windows 7 64-bit previously but I completely wiped it). Any ideas?

    Read the article

  • Will the VMs on the ESXi cluster be running after disconnection from the network?

    - by John
    What will happen if I disconnect the ESXi cluster configured with HA from the switch(I need to change the power source on the main switch) and there is no management port redundancy? I'm going to disable the host monitoring and VM monitoring within HA settings, and connect to the switch after it boots up. Will the virtual machines be running if I disconnect the hosts from the network so they will not see any other ESXi host? I hope everything will work fine and the hosts will join the cluster after they connect to the network again, but I would like to be sure ..

    Read the article

  • What does the red x icon mean next to a user in folder permissions (Windows 7)

    - by Scott Szretter
    In trying to debug various strange issues on a machine, I found something strange - when I go to C:\Users\administrator and get properties, security tab, it lists the users (the local admin account, system, and 'administrator' which is the domain administrator account). It all looks fine in terms of permissions (full control, etc.) compared to other machines. The one difference is there is a small red circle with an X to the left of the user icon/name. Additionally, there are various folders where it says access denied under there - for example, my documents! Even logged in as the local machine administrator account (which is not named administrator), I am unable to change the permissions - it says access denied. Any ideas what this means and how to fix it? I even tried re-joining the machine to the domain.

    Read the article

  • Website is live but ping times out

    - by infinity
    I have a client's website that is running on GoDaddy hosting and started behaving very strange recently. The site is up and running but when I try to ping it I get time out. The problem is PayPal doesn't work also Google Webmaster reported the site is down. The client spoke with support and they said there is no firewall or any other traffic filter on their side. The site itself is PHP. Any ideas are welcome. I've tried to send test IPN from PayPal to the payment URL and got: IPN delivery failed. Unable to connect to the specified URL. Please verify the URL and try again. Which makes me think that the website/server is inaccessible in some specific cases. The site URL is http://www.flavourly.com P.S.: I tried to ping it from different machines, ISPs and OS

    Read the article

  • External disk not working with Mac OS X 10.5

    - by ctx2002
    Hi: My OS is 10.5.7. I have an external hard disk, formatted as Mac OS Extended (Journaled), connected to my Mac machine through USB cable. I have used it for a while, but now it is not working. Error showed in /var/log/system.log. USBF: 2521.384 AppleUSBEHCI[0xa515800]::Found a transaction which hasn't moved in 5 seconds on bus 0xfd, timing out! (Addr: 0, EP: 0) ..... ...... kernel[0]: USBF: 2588.616 [0xa6baa00] The IOUSBFamily was not able to enumerate a device. Luckily, in my work place there is a Mac with OS 10.4, so I had plugged the disk into the 10.4 machine and it worked!!! I did checked "System Profiler" on both machines (10.5, 10.4). Under usb section, for 10.5 it shows USBEHCI, under 10.4, it shows USBOHCI. I am not a expert on USB Device , can any one tell me why 10.5 cause the problem. it was worked under 10.5

    Read the article

  • Sharing a symlinked (`mklink /d`) directory via SMB?

    - by Alois Mahdal
    I have a Windows 7 amd64 box where one directory is shared: local path is d:\drop\ remote path is \\aloism\drop from SMB point of view, Everyone has Read and Write permission ACLs for the folder are set so that all authenticated users have read and write permissions:NT AUTHORITY\Authenticated Users:(OI)(CI)C (which is inherited to all levels below) Now I create a symbolic link within the structure of the directory: D:\drop>mklink /d tools2 tools symbolic link created for tools2 <<===>> tools The problem is that I can't access this new directory from any of the remote machines (a Windows 7 box and a Windows XP box—both behave the same way): C:\>dir \\aloism\drop\tools2\ Volume in drive \\aloism\drop is droot Volume Serial Number is FA73-1897 Directory of \\aloism\drop\tools2 File Not Found How can I make it work? Possibly also for files?

    Read the article

  • Multiple PCs - Single profile

    - by martixy
    Or how to create an AD(or similar) environment at home? So... I work in this big IT company(they have and AD environment ofc) and I've seen first hand how nice integration and log-in-from-anywhere(roaming?) profiles can be. So I wanna create the same setup at home, where I have 3 machines. Mostly logging in with the same profile and preserving the settings(things like UAC being off, a few group policies I've customized, things like that). How can I achieve something like this?

    Read the article

  • SSH tunnel doesn't work

    - by s1ck
    I am trying to use my server as a "proxy" with ssh. However, setting up tunneling with ssh -D localhost:8000 user@myserver does not work. I tested this on various machines with ssh and putty - It connects just fine, but when I set my browser settings accordingly, I just get an error "Connection has been reset". I tried monitoring the traffic with wireshark, but I didn't even see some tunnel-traffic. I explicitly set AllowTcpForwarding to "yes" but I still can't use the tunnel. When running ssh in verbose mode, I don't get any errors but debug1: Connection to port 8000 forwarding to socks port 0 requested. debug1: channel 3: new [dynamic-tcpip] debug1: channel 3: free: dynamic-tcpip, nchannels 4 What am I doing wrong?

    Read the article

  • How to make ssh match known_hosts to host/ip:port instead of just host/ip?

    - by Prody
    I have two machines behind a firewall, with the ssh ports forwarded to 2201 and 2202. When I ssh host -p 2201 it asks if I trust the machine, I say yes, it gets added to ~/.ssh/known_hosts Then I ssh host -p 2202 it doesn't let me, because there's already a known_host for this IP in ~/.ssh/known_host:1 (the file was empty when I started, so line 1 is the one added by the previous ssh run) This happens on CentOS 5.4. On other distros (I've tried Arch), it appears that ssh matches the knwown_hosts to the ports too, so I can have multiple fingerprints for multiple ports on the same host/ip without any problems. How can I get this same behavior for CentOS? I couldn't find anything in man ssh_config. (or at least not without disabling fingerprint checking)

    Read the article

  • Is it safe to sync a Firefox instance in both directions?

    - by java.is.for.desktop
    Hello, everyone! I want to use some tool (not decided yet, which one) to sync a Firefox instance (more exact: user directory) between two machines. (EDIT: I want really to sync everything in the user-directory) Would you assume, that this is safe? What would happen, if some files are newer on the one machine, and some other files are newer on the other machine? Could this lead to inconsistencies, when, let's say, there are some inter-file references? If general, I don't have good experience with syncing application in both directions. Most applications seem not to be suited for this.

    Read the article

  • Cannot login SQL Server after changing machine name

    - by Ucodia
    After installing and setting up new machines in a domain, we decided to rename one of them which had a SQL Server instance installed. So I changed the hostname, everything went fine regarding the domain but now, the server is logging a approximatively 2 SQL Server errors every 5 minutes and I cannot connect to the instance localy or from anywhere within the domain. Here is the error from the event log: SSPI handshake failed with error code 0x8009030c, state 14 while establishing a connection with integrated security; the connection has been closed. Reason: AcceptSecurityContext failed. The Windows error code indicates the cause of failure. The logon attempt failed [CLIENT: x.x.x.x] Concerning the instance, everything is started and restarted without any extra error.

    Read the article

  • Minimal Lync 2013 Setup

    - by vkapadia
    I would like to install a Lync 2013 server at home. What would be the minimum requirements for it? I've seen the requirements and they seem pretty hefty, just want to know what I need for a basic set up. This will be for up to 5 users. I would like users to be able to connect from outside the network as well. From what I've been able to find, it looks like I'll need 3 servers. A domain controller, a lync front end server, and an edge server (for external connectivity, and it will need 2 network cards). Is that really all necessary? And how would they need to be connected (I've never connected a machine with multiple network cards)? Do they all have to be separate machines, or can I use VMs?

    Read the article

  • Best practices for FQDN for standalone domain (is a two part domain.tld okay?)

    - by birchbark
    I've searched quite a bit and can't seem to find a straight, modern answer on this. If I am hosting a domain, say, mydomain.com, on a machine which is going to solely be used for that domain, and there are no subdomains, is there a real, practical reason besides compliance to create an arbitrary hostname (i.e. myhost) just in order to have a three-part FQDN (myhost.mydomain.com) to satisfy some sort of RFC or convention that's expected. This seems to make a lot of undue complexities from my perspective, and I'm not sure if there's an advantage to this or if it's just a hold-over from a time where all web resources came from a subdomains such as www and ftp which may need to scale to separate machines. I don't use www on my domain, either, which is ill-advised for all I know from an administrators perspective (though removing it is the norm from a designer's perspective)...

    Read the article

  • How should a small team using multiple OS's deploy over github?

    - by Toby
    We have a small development team that have recently moved to using github to host our projects. The team consists of three developers, 2 on Windows and 1 on Mac. I am currently researching the best way to deploy applications to our Linux servers (dev and production). Capistrano running locally would be ideal but from what I read this won't work for Windows machines. It looks like the best way is to use a post-receive hook in github, I can see how this would work for auto deploying to dev, but I don't see how we could then deploy to live. I have found paid projects like http://www.deployhq.com/ but it feels like something that a quick bit of code should be able to do for free, I just can't seem to get myself pointed in the right direction! I was wondering what would be considered best practice for small team deployment involving multiple local OS's and github.

    Read the article

  • Remotely initiate windows updates

    - by TetraFlash
    I have a network with countless workstations on it and the windows updates are done through WSUS in push cycles to allow us to use a generic image and configuration and avoid network congestion. I have a number of workstations in storage that are connected once a month for a few days to recieve updates, however not all of them fall within their push cycle. I want to avoid going to each machine and clicking "check updates" as that would require me to connect a monitor, mouse and keyboard. Is there a way (preferably through powershell but im open minded) to initiate a check and install of updates on a remote system? Lets say for 1 system at a time right now, I can add a file reader and a look later. NOTE: the configuration of the WSUS server is above my pay grade, im simply being asked to ensure these machines are updated. Thanks for any help or hints

    Read the article

  • IIS 6.0 :: SSL certificate install :: issue

    - by BSI Support
    Certificate request process works fine. Once the certificate is received from VeriSign, the installation process also works fine (no errors.) However, once installed, the new certificate is not "assigned." Meaning, when I view the IIS instance security properties, it still shows the previous certificate start & end dates. Thoughts? Windows 2003, Web Edition. I have 3 other Win2003WE machines, configured nearly the same, and do not have this problem.

    Read the article

  • Is it possible as an Administrator to gain access to a SQL Server 2008 instance without changing any passwords?

    - by adhocgeek
    I have administrative access on our network, but I don't manage the installation of all servers or software. On some of our machines instances of SQL Server 2008 have been installed which I need to be able to access, but since my account hasn't been explicitly granted a login, I can't get into. Is there a way to get into the database without changing anyone's password (e.g. I could solve this by changing the password of the user who installed the instance, assuming they've set themselves up as admin, and then logging on as them, but I don't want to have to do this).

    Read the article

  • How do I log in to the ubuntu 10.04 server using grub2 from a remote computer/terminal with GUI running

    - by Steve Cornall
    I upgraded my server from ubuntu 8.04 to 10.04. It now uses the grub2 bootloader. In 8.04 from the grub log-in page I could select the option log in to server and thereby connect through SSH directly to my server and have the GUI running. I can currently log in to my upgraded server now from any of my 8.04 machines using grub1 with the GUI running. I cannot log in that way from a machine that uses grub2 and ubuntu 10.04. I want to upgrade my entire network to ubuntu 10.04 but cannot until I know how to log in to the network from grub2 with the gui open. I have exhausted all my ideas as to the solution to this problem. Any help would be most appreciated. Thank-you

    Read the article

  • Setting up central git repo on local Mac network

    - by Dashman
    We are a team of three, all working on our local machines on the same internal network. We will all be working on websites in local working copies of the same Git repo hosted on Github. We have an internal staging machine here (dev.internal), and I am looking for a way for us to be able to push to this machine. At each milestone in the development cycle. In essence, all I really want us to be able to do is add the dev.internal machine as a remote, and push to this whenever we are ready. Could somebody please point me in the right direction to get this set up?

    Read the article

  • Use puppet to make changes to ip route and sysctl

    - by Quintin Par
    I have two changes to ip route & sysctl that disable tcp slow start. Here’s how I do it ip route show Make a note of the line starting with default. Pick up the IP from the default line and run sudo ip route change default via $ip_address dev eth0 initcwnd 12 sudo sysctl -w net.ipv4.tcp_slow_start_after_idle=0 How can I create a puppet script out of this? One that can be deployed to many machines of the same type – CentOS 6 Edit: Added bounty to get a working example for sudo ip route change default via $ip_address dev eth0 initcwnd 12

    Read the article

  • ssl problems between two tomcat servers

    - by user1058410
    I've got one tomcat 7 server(kind of like a web server) trying to talk to a tomcat 6 server(acting as a document server on another machine) over ssl and keep getting this error java.security.cert.CertificateException: No name matching rarity64 found. Where rarity64 is the name of the document server. I've tried exporting keys from both tomcats keystores and importing them into the others keystores using java keytool. I've even tried adding them to the other machines cacerts keystore. I've also used internet explorer to import both keys into the other machine. But nothing I try works. If it matters the real webserver is IIS 7.5, which the tomcat "webserver" talks too with arr, and they don't use SSL. But the problem seems to between the two tomcat servers

    Read the article

  • Recommendation for hardware upgrade: thin clients? Or...?

    - by Alex C.
    I work for an animal shelter in Upstate New York. We have about 50 machines running XP Pro. They're connected to a Windows network with a domain. About half of these computers are used for nothing more than using two web-based apps -- one to keep track of our animals, the other to process credit cards. Having a full-blown desktop PC seems like overkill for this purpose. The PCs are three-to-five years old, and I'd like to come up with a plan to upgrade the hardware. Our donations are down (not surprising, given the economy), so cost is a big factor. Can people recommend some options? Some sort of thin client, maybe?

    Read the article

  • Previous Versions delayed with SBS 2008 / Windows 7

    - by indeed005
    The SBS 2008 at this site has Previous Versions enabled on a mapped drive. The snapshots happen 3 times a day. I doubt they take very long; the diff size is only a few GB. The problem is that the users on Windows 7 cannot see their previous versions until a few hours later. Is there some indexing that has to happen before the Previous Versions are visible to the client machines? Edit: In the application log, 3 minutes after every scheduled backup, there is an event for VSS (EventID 8224) saying "The VSS service is shutting down due to idle timeout" Apparently this means it has finished successfully, even though modified files still do not show another version.

    Read the article

  • Internet connection fails in Ubuntu on VirtualBox when virtual machine is created from "Import appli

    - by Sanoj
    I have installed Ubuntu Server 9.10 in a virtual machine in VirtualBox, then I made a cope/clone and exported it with "Export appliance" so I can create many cloned virtual machines. But when I try to import an appliance, everthing seams to be fine with the Ubuntu except that it can't connect to Internet and doesn't get an IP-address. The machine is used in Bridged mode. And it doesn't help to change to NAT-mode either. The machine that I cloned seams to work fine, and get an IP address. How to fix this? Where am I doing wrong?

    Read the article

  • Testifying that somaxconn change can make a difference

    - by petermolnar
    I got into an argument on the net.core.somaxconn parameter: I was told that it will not make any difference if we change the default 128. I believed this might be enough proof: "If the backlog argument is greater than the value in /proc/sys/net/core/somaxconn, then it is silently truncated to that value" http://linux.die.net/man/2/listen but it's not. So does anyone know a method to testify this with two machines, one running, for example MySQL or LVS and the other is hammering it in a GBit network? I'm opened to any solution, scripts are slightly more welcome.

    Read the article

< Previous Page | 139 140 141 142 143 144 145 146 147 148 149 150  | Next Page >