Search Results

Search found 2694 results on 108 pages for 'michael ellick ang'.

Page 44/108 | < Previous Page | 40 41 42 43 44 45 46 47 48 49 50 51  | Next Page >

  • How can I use Android as a remote control for streaming?

    - by michael
    I would like to know if I can use Android to remote control streaming on my laptop? I would like to use my laptop as my streaming server and use my HDTV to view the stream. And I need some way to remote control my streaming server. I have read about http://maketecheasier.com/install-vlc-shares-in-Ubuntu-and-stream-videos-to-Android/2011/02/25 and http://code.google.com/p/android-vlc-remote/ but those are streaming to Android phone itself. I am just need something to remote control streaming to my TV. Is that possible?

    Read the article

  • IIS 7: One Page Works, All Others Fail With "Error code: ssl_error_rx_record_too_long"

    - by Michael
    On my local machine, I have a second site bound to Port 81. Within that site is a certain page which I can browse from other machines with no problems, but all other pages fail with "Error code: ssl_error_rx_record_too_long". Each of the failing pages (as well as the lone working page), works with localhost. So, from any machine, local or remote: http://cmwmach01.mydomain.biz:81/RD/SS/SS.aspx (works) http://cmwmach01:81/RD/SS/SS.aspx (works) http://cmwmach01.mydomain.biz:81/RD/POV/SC.aspx (fails - gets changed to https) http://cmwmach01:81/RD/POV/SC.aspx (fails - gets changed to https) Everything works with localhost (locally, of course). I've tagged this question with SSL because, at one point, it would warn about an SSL cert issue (maybe this was self-signed at one point?), but now it doesn't. While there may be an issue around that, I don't see how this could cause the issue I am seeing (but, as I mention below, am I way out of my depth here). I am way out of depth here in trying to figure out why that one page works (or the others don't), so that I can make them all work. Any ideas?

    Read the article

  • How do I open files from the command line in Mac

    - by Michael
    I'm following a video tutorial where the author (who uses textmate) can open files by using "mate". for example mate .git/config will open this config file I'm using textwrangler however so I don't have that option. I did try edit .README once when i tried to open the README file of an application, but it opened a blank README file in textwrangler instead of the file with the text in it so any idea how I can open this .git/config file (or any other file) using textwrangler? I'm using Mac snow leopard

    Read the article

  • Firefox extension to translate selection boxes

    - by Michael
    Many extensions can translate selected texts. What if the text located within selection box? Meanwhile I can't translate the whole website because google translator says "Sorry, we are unable to translate the page you requested." I can use firebug to explore html, copy the text to google translator and do translation. Wondering if there is such add-on for firefox to translate text from "non selectable" HTML elements. Thx

    Read the article

  • Any software to remove $NtUninstallxxxxx

    - by Michael
    Any commercial or free software to give me a list and descriptions of patch uninstalls and let me remove selected ones? I've tried Windows XP Update Remover, but seems it doesn't provide any information for majority of items and I have to delete one by one... I also know I can do it manually, but just wondering if there is more professional software to make it more accurate and quick.

    Read the article

  • Need help in using svn on ubuntu 9.10

    - by michael
    Hi, I have install svn on ubuntu 9.10. But when I try to use svn to checkout code for an open source project, i get this error: $ svn co svn://svn.valgrind.org/valgrind/trunk valgrind svn: Berkeley DB error for filesystem '/home/svn/repos/valgrind/db' while opening 'nodes' table: Cannot allocate memory svn: bdb: Lock table is out of available locker entries Can you please tell me how to fix it? Thank you.

    Read the article

  • Any reasonable UPS for a Desktop PC, just to shut it down?

    - by Michael Stum
    While I do have a surge protector to protect against overvoltage (hopefully), I have nothing against undervoltage. When a lightning storm hits, I had the lights flickering at some point. The PC continued to run, but it got me thinking of getting a UPS as a way to a) have a clean 120V/60Hz power source and b) have a way to shut down the PC in case something bad happens. I heard not all UPS' protect against power spikes, so I wonder if someone has a recommendation? It does not need to keep the PC on for a long time if the power goes out, it's good enough if it shuts down the PC after 5 minutes or so. There are 2 PCs connected. One is a Core i7-860 with a Radeon 5870 running Windows 7 Ultimate (so quite power hungry. It uses a 600W PSU but I have no measurements of the actual usage), the other one is a Windoes Home Server, running WHS/Windows Server 2003. Any recommendations in the low-price segment?

    Read the article

  • Exchange 2010 SP1 upgrade prerequisite checks fail for KB979744, but update fails as Not Applicable

    - by Michael Shimmins
    I'm trying to install the SP1 update to an RTM Exchange 2010 box. The Client Access Role Prerequisites are failing, complaining that KB979744 isn't installed. I downloaded the update and tried to install it, however it will not install complaining that 'The update does not apply to your system'. A discussion on TechNet indicates that I need .NET 2.0 or 3.5 installed (both are). Aside from that I've been unable to find anything helpful. Hopefully someone has encountered this and knows how to get around it, don't really want to use up a support call on this. Edit: This is on a 64-bit edition of Windows Server 2008 Datacenter (without Hyper-V) SP2.

    Read the article

  • How to avoid Remove-Item PowerShell errors "process cannot access the file"?

    - by Michael Freidgeim
    We are using TfsDeployer and PowerShell script to remove the folders ising Remove-Item before deployment of a new version. Sometimes the PS script failed with the error Remove-Item : Cannot remove item Services\bin: The process cannot access the file Services\bin' because it is being used by another proc Get-ChildItem -Path $Destination -Recurse | Remove-Item <<<< -force -recurse + CategoryInfo : WriteError: (C:\Program File..\Services\bin:DirectoryInfo) [Remove-Item], IOException FullyQualifiedErrorId : RemoveFileSystemItemIOError,Microsoft.PowerShell.Commands.RemoveItemCommand I’ve tried to follow the answer from PowerShell remove force to pipe get-childitem -recurse into remove-item. get-childitem * -include *.csv -recurse | remove-item ,but the error still happens periodically. We are using unlocker to manually kill locking application, (it’s usually w3wp), but I prefer to find automated solution. Another (not ideal) option is to-suppress-powershell-errors get-childitem -recurse -force -erroraction silentlycontinue Any suggestions are welcome.

    Read the article

  • How to get linux to stop asking "There are # rows, list them anyway? [n/y]" for auto-completing

    - by Michael
    How do I get tcsh to stop asking if I want to list files in a directory that may have a lot of auto-completes? For example, if I do: xemacs ../"TAB" to get the list of files it asks: There are 371 rows, list them anyway? [n/y] I don't want it to ask this, just list them.. it's getting tiresome. (Or how to at least set the tolerance to a higher number of auto-completes before it asks) Thanks.

    Read the article

  • Backup all plesk MySQL Databases to individual files

    - by Michael
    Hy, Because I'm new to shell scripting I need a hand. I currently backup all mydatabases to a single file, thing that makes the restore preaty hard. The second problem that my MySQL password dosen't work because of a Plesk bug and i get the password from "/etc/psa/.psa.shadow". Here is the code that I use to backup all my databases to a single file. mysqldump -uadmin -p`cat /etc/psa/.psa.shadow` --all-databases | bzip2 -c > /root/21.10.2013.sql.bz2 I found some scripts on the web that backup each database to individual files but I don't know how to make them work for my situation. Here is a example script: for db in $(mysql -e 'show databases' -s --skip-column-names); do mysqldump $db | gzip > "/backups/mysqldump-$(hostname)-$db-$(date +%Y-%m-%d-%H.%M.%S).gz"; done Can someone help me make the script above work for my situation? Requirements: Backup each database to individual file using plesk password location.

    Read the article

  • How to launch Google Chrome Application Shortcuts in Linux?

    - by Michael Rose
    I've got Chrome running on my Linux netbook, it's great and, unlike the Mac version, the 'Create Application Shortcut' option isn't greyed out. So I created one for Gmail. The 'applications' get stored at ~/.local/share/applications but I haven't managed to launch it yet. If I use Chrome in terminal $ google-chrome ~/.local/share/applications/google-application-reallylong-randomname.thing It opens the file in a new tab in Chrome and helpfully downloads it for me! Anyone got application shortcuts working in Linux?

    Read the article

  • Lighttpd based server issues crop up when port forwarding

    - by michael
    I have four host computers running lighttpd webservers. they are sitting behind a hspa modem, which each occupying a http port between [81 - 84]. 80 is taken by the modem itself. The port forwarding is setup correctly, however, only a portion of any webpage I request from any of the hosts comes through (they all fails after %20 of the page). If I put the host on port 81 into the dmz, it serves pages fine. The others do not respond to the dmz treatment. Is it possible the web content on the hosts somehow require ports aside from their respective http port? Or is it possible that even though the server.port in the lighttpd_ssl.conf file is set, the individual hosts are still expecting to serve on port 80? I am not familiar with lighttpd, nor did i set them up. they are running on video encoders i purchased. I can grab any files from them required for further information on the problem.

    Read the article

  • Alternative to Softerra's LDAP Administrator?

    - by Michael Mao
    I am looking for an alternative to Softerra's LDAP Administrator, simply for the price is over-budget for a small team to test the LDAP credentials. So I really appreciate if someone can point me to some other alternatives for a Windows-based environment. We've got our Active Directory setup already, I only need a tool to confirm the access as a proof to customers. Sorry I am not a security guru, I am quite the opposite... Thanks for any suggestion in advance!

    Read the article

  • What special configuration is needed to have a domain controller with DNS on a VM in Hyper-V?

    - by Michael J. Gray
    I have a server running Windows Server 2012 and it only has the Hyper-V role. I have a VM also running Windows Server 2012 and it has the Active Directory: Directory Services and DNS ([example.com]) roles. It has been promoted to a domain controller for [example.com]. In the virtualization host I have joined the domain successfully. On a laptop on the network, it can ping my DNS that is inside the VM on the virtualization host. However, it can't resolve anything through the DNS. I figured it was a firewall issue where the DNS had the firewall allow rule, but the virtualization host did not. I added that, but it didn't fix the issue. On my laptop, I get "UnKnown can't find [example.com]: No response from server" immediately. There's no time out, it just comes back right away. What have I overlooked?

    Read the article

  • What's a worthwhile test for a new HD?

    - by Michael Kohne
    I work for a company that uses standard 2.5" SATA HD's in our product. We presently test them by running the Linux 'badblocks -w' command on them when we get them - but they are 160 gig drives, so that takes like 5 hours (we boot parted magic onto a PC to do the scan). We don't actually build that many systems at a time, so this doable, but seriously annoying. Is there any research or anecdotal evidence on what a good incoming test for a hard drive should be? I'm thinking that we should just wipe them with all zeros, write out our image, and do a full drive read back. That would end up being only about 1 hour 45 minutes total. Given that drives do block remapping on their own, would what I've proposed show up any infant mortality just as well as running badblocks?

    Read the article

< Previous Page | 40 41 42 43 44 45 46 47 48 49 50 51  | Next Page >