Search Results

Search found 68249 results on 2730 pages for 'sudo work'.

Page 76/2730 | < Previous Page | 72 73 74 75 76 77 78 79 80 81 82 83  | Next Page >

  • Can't get .htaccess to work

    - by orokusaki
    I'm using Apache2 on Ubuntu Lucid Lynx. I have config set to use .htaccess like normal. This is my default site: <VirtualHost *:80> ServerAdmin webmaster@localhost DocumentRoot /var/www <Directory /> Options FollowSymLinks AllowOverride All </Directory> <Directory /var/www/> Options Indexes FollowSymLinks MultiViews AllowOverride All Order allow,deny allow from all </Directory> ScriptAlias /cgi-bin/ /usr/lib/cgi-bin/ <Directory "/usr/lib/cgi-bin"> AllowOverride All Options +ExecCGI -MultiViews +SymLinksIfOwnerMatch Order allow,deny Allow from all </Directory> ErrorLog /var/log/apache2/error.log # Possible values include: debug, info, notice, warn, error, crit, # alert, emerg. LogLevel warn CustomLog /var/log/apache2/access.log combined Alias /doc/ "/usr/share/doc/" <Directory "/usr/share/doc/"> Options Indexes MultiViews FollowSymLinks AllowOverride None Order deny,allow Deny from all Allow from 127.0.0.0/255.0.0.0 ::1/128 </Directory> </VirtualHost> I've tried lower case "all" (AllowOverride all) as well. My .htaccess file looks like this: //Rewrite all requests to www Options +FollowSymLinks RewriteEngine on RewriteCond %{HTTP_HOST} ^mydomain.com [nc] RewriteRule ^(.*)$ http://www.mydomain.com/$1 [r=301,nc] //301 Redirect "old_junk.html" File to "new_junk.html" Redirect 301 /old_junk.html /new_junk.html //301 Redirect Entire Directory "old_junk/" to "new_junk/" RedirectMatch 301 /old_junk/(.*) /new_junk//$1 // Copy and paste redirect examples from above: (with mydomain replaced with my actual domain... and my computer is plugged in)

    Read the article

  • LFTP when used with proxies doesn't work

    - by user2949465
    can't seem to use LFTP with proxies that require authentification correctly on my Ubuntu server. When I use it with proxy that doesn't require username/password everything seems fine: lftp lftp :~> set http:proxy http://HOST:PORT lftp :~> set ftp:proxy http://HOST:PORT lftp :~> open username:[email protected] lftp [email protected]:~> get file.ext file.ext 36352 bytes transferred in 10 seconds (3.5K/s) lftp [email protected]:~> exit but when I have to put username/password there is a problem: lftp lftp :~> set http:proxy http://proxylogin:proxypass@HOST:port lftp :~> set ftp:proxy http://proxylogin:proxypass@HOST:port lftp :~> open ftp://ftpuser:[email protected] answer: cd: Access failed: 401 Authentication Required (~) please someone help!

    Read the article

  • XP Mode in Windows 7 professional recently ceased to work

    - by Ian FLetcher
    I have used Windows XP Mode in Windows Professional 7 once a week for around 2 years to run Quicken 98. About 2 weeks ago the XP Mode completely stopped working. I get different responses when I try to launch. I can sometimes see a launch screen for XP (which I never used to get) which then hangs Sometimes I just get a black screen Sometimes I get an error message in a cmd like interface that says "windows could not start because the following file is missing or corrupt \system32\ntoskrnl.exe please reinstall a copy of the above file. I have noticed that in the tools tab at the top of the screen "enable integration features" is greyed out. Any recommendations? Ian

    Read the article

  • How to legitimately work around ISP rate limits

    - by Derek Ting
    A lot of ISP rate limit the amount of e-mails that is sent from a particular IP address. What is the proper way to get around that rate limit? Our company has an iPhone application that sends many e-mails because of our large user base and many e-mails go to different ISPs that rate limit the number of messages coming from a specific IP. We do not send spam and we are a legitimate business. However, is there a better way to resolve this limitation rather than just getting a ton of IP addresses? Ideally, I wouldn't want to rely on a third party service to send mail. However, if its the only possible solution, we would consider.

    Read the article

  • Automated git push attempt does not work - authentication issue

    - by at least three characters
    I'm trying to automate a very periodic git add/commit/push cycle using a shell script and cron under OS X 10.8.5. The script is as basic as one would expect it to be: cd /my/directory git add . git commit -m "a commit message with the date" git push -u origin master I've tried running it both as root as well as a non-root user. When I do this manually, I get a dialog box from OS X requesting that I authenticate the operation. Running the script (either using cron or just using sh) ends up sending a message (via mail) to whichever user's cron executed the script saying that it was unable to write a file in the .git directory because of a permissions issue (which is most likely manual execution requires authentication). Is there any way to circumvent this issue, or give the script permission to perform this operation without having me intervene each time?

    Read the article

  • Nameservers Won't Work

    - by user39110
    Hi, i have dedicated server and i use vendabilisim.com for nameserver. i have 2 nameservers ns1.vendabilisim.com 213.128.64.91 and ns2.vendabilisim.com 213.128.64.92 ip addresses are working but nameservers don't. Please help.

    Read the article

  • Git not work on my Mac OS X .

    - by koko
    I downloaded and install Git from http://code.google.com/p/git-osx-installer/ . After installing I can't use my Git installation. thomas-macbook-:test zozo$ git init -bash: git: command not found What is wrong? Many thanks in advance.

    Read the article

  • Fedora Installation with software repository in DVD does not work

    - by Raks
    I bought a new assembled PC with processor as Core-i3(2120) and Intel H61 motherboard and was trying to install Fedora 16 from a DVD. This DVD contains all the packages so that installation does not require to download packages from internet. I have used this DVD to install Fedora 16 offline many a times though in machines with different hardware configuration. But in this new machine when the installation reaches the stage wherein it asks for Software repository selection I select CD/DVD but the system fails to read the media and throws up an error that it cannot detect the media. THe LED in the DVD writer also indicate that the DVD is not being read. Now there is neither a problem with the DVD or the DVD drive because the installation started from the DVD itself. So what could the problem be, anything in the BIOS that is causing the problem, Is there any way I could utilize the packages already existing in the CD so that I save downloading the packages from DVD

    Read the article

  • Can't get bridged networking to work between linux guest virtual machine and Mac host

    - by tgoneil
    I'm trying to establish bridged networking from linux mint vsn 12 in virtualbox to a Mac Lion host. Mac config: Network setting: en3 configured by DHCP Sharing setting: Internet Sharing selected, Share connection from en3 to computers using en3 Virtualbox Linux setting: Network setting: Bridged Adapter, Name: en3 I can ping from host (192.168.2.1) to guest (192.168.2.2) and guest to host, but I Cannot ping from the Linux guest to the outside world. Connection in host is up, because I can ping from the Mac host to the outside world. Something else that's seems weird to me, in the Mac Network setting, the IP Address generated by DHCP says 169.254.243.185. What the heck is that?? When I open a terminal up in the Mac, however, ifconfig shows its en3 inet address as 192.168.2.1.

    Read the article

  • Eclipse Pydev Ctrl-Click (Go to Definition) Doesn't Work OSX

    - by Koobz
    My Pydev set up in OS X is kind of busted. I'm working on a Django project and I find that Ctrl-Click never actually goes to the definitions of any of my objects or functions. I actually have a symlink to Django/django in my workspace so that it's easier to cross reference Django code. My guess is that something is wrong with the builder, but it doesn't throw up any errors. Does anyone have advice here? Different topic: does anyone know of a good way to use Ctrl-Shift-R (Open resource) and filter files by folder? It's not that useful in python projects where you have 20 urls.py showing up.

    Read the article

  • Installation of Active Directory on separate VM from DNS does not entierly work - not sure why

    - by René Kåbis
    Not sure what I am doing wrong here. I have a moderately midrange server (16 cores, 2Ghz, 32GB ECC REG RAM, 6TB storage, nothing too extreme) where I am running Hyper-V (Server 2012 R2 Enterprise) in order to provision virtual machines. So why an AD separate from DNS? I want redundancy. I want to be able to move VMs and back them up individually and not have too many services on any one VM. I have already provisioned a VM with DNS, and have set it up right -- essentially, I have: Set up Static IP’s for everyone involved. Installed the DNS service on the DNS VM. Created a forward lookup zone and a reverse lookup zone (primary zone) xyz.ca Configured the zones to use nonsecure and secure dynamic updates (i will change this to secure later after the domain controller is online). Created a A record for the DC in the forward lookup zone (and a reverse ptr) Changed DC’s DNS server (network settings) to the new DNS server. Checked that I can ping the dns server from the new DC by hostname. When I went ahead and did a DCpromo on the DC, and un-cheked the “install DNS” option, everything seemed to go well (no error messages), but I saw no changes on the DNS server whatsoever (no additional settings). Plus, the DNS server seems to be unable to join the domain, as it claims that the domain is not discoverable. As a final note, I do run Symantec Endpoint Protection, which includes a firewall and most settings set as default. I have not yet tried turning this off, but my experience has been that if a service would open up a port on a Windows firewall, it would do the same through Symantec. There is pretty tight integration these days with corporate-class AV and Windows. I have a template vhdx fully set up (just short of any special roles and features) that I can use to replace the current AD VM with, so doing this all over again is not too much skin off of my nose.

    Read the article

  • Network Explorer Intermittently Fails to Display all Computers in Work Group

    - by graf_ignotiev
    I run a small computer lab of 10 computers and occasionally, when using the network explorer (a.k.a Network Browser) some or all of the remote computers will fail to appear. If I try to access a remote computer by its name I get an unspecified error (code 0x80004005), but I am still able to access it with the computer's IP address. The strangest part is that the problem will inexplicably go away after waiting awhile. Each computer is running Windows 7 x64 Enterprise and has identical hardware, software and configuration. They are all on the same subnet and in the same workgroup. I've spent days researching the problem and have tried the following solutions: Updated the BIOS, chipset and network adapter drivers Changed Power Settings in Network Adapter Properties so that the computer will not turn it off Disabled the Computer Browser service Changed the DHCP node type to broadcast Reviewed the Event Viewer logs Steps 3 and 4 have seemed to help the problem a little bit, but not completely. I'm beginning to suspect that the problem might lie with our router which is a ZyXEL ZyWALL 2WG, as the packets sent by Network Discovery may not be returning in time, but I wanted to get some perspective in the issue before I went any further.

    Read the article

  • Normal check interval and re try check interval doesnot work in nagios

    - by chandra
    Hi Guys, I have configured nagios in my environment and i have used the following setting for a service which monitors the disk space for every 10 min , Normal check interval =10 min retry check interval = 1 min Maximum attempts =2 But accoring to this values i should recieve my first alert after 10 min and then from there for every 1 min. But in my case i am recieving my first alert and then next alert comes to me after 20 min. Can some one help me in fixing this.

    Read the article

  • Make cloudera-vm work on Oracle VM VirtualBox

    - by ????? ????????
    I downloaded this and the instructions say: Important: You must enable the I/O APIC in order to use 64-bit mode. (See http://www.virtualbox.org/manual/ch03.html.) On newer versions of VirtualBox, it may default to using SATA as the disk interface. This can cause a kernel panic in the VM. Switching to the IDE driver solves this problem. I am running this on Red Hat 64-bit mode (I've also tried on Ubuntu 64-bit with the same result). I pointed to the cloudera-vm image as a startup disk for the VM. I am getting this message: Failed to open a session for the virtual machine ClouderaDevelopment. VT-x features locked or unavailable in MSR. (VERR_VMX_MSR_LOCKED_OR_DISABLED). Result Code: E_FAIL (0x80004005) Component: Console Interface: IConsole {1968b7d3-e3bf-4ceb-99e0-cb7c913317bb} Does anyone know what I am doing wrong?

    Read the article

  • Why does my allow_url_include not work?

    - by autthapone
    Server Information: CentOS 5.7 (Final), PHP Version 5.2.6, Apache/2.2.3 I edit in /etc/php.ini change to allow_url_include = On. Then restart apache. I see configuration on phpinfo() file, but allow_url_include not changed. It's Off yet. Help me, please. My Setting. - http://postimage.org/image/aliuyb9a3/ My phpinfo - http://postimage.org/image/tlsu18b1h/ I can't find other php.ini file. upload_max_filesize also not changed :-( but max_execution_time and memory_limit is changed. Everyone, if issue can't solving now, please guide me about repair/re-install PHP on CentOS.

    Read the article

  • None of my bash commands work

    - by Kevin
    I have an Ubuntu 9.10 netbook. I has always run great. Two days ago, I was running as root for a while (~30), and when I moved back to my user account (only other account one this machine), all the commands in ~/bin stopped working. If I try ls, it comes up with "cannot execute binary file". Same with ln, mv, mkdir, clear, cp, etc. They all run as root(which makes sense, different files), but I have no idea why this happened. I don't want to stay as root to move around easily. Any idea?

    Read the article

  • Uninstaller that can work with failed installs

    - by Nifle
    I'd like an application that can uninstall applications that does not show up in the normal add/remove applications. I have an application that failed to install properly. It installs half way and then crashes leaving short cuts, folders with exe and config files in and some stuff in the registry. Ideally I would like to be able to point the uninstaller to the install dir and tell it to remove everything that references files in that dir.

    Read the article

  • None of my bash commands work

    - by Kevin
    I have an Ubuntu 9.10 netbook. I has always run great. Two days ago, I was running as root for a while (~30), and when I moved back to my user account (only other account one this machine), all the commands in ~/bin stopped working. If I try ls, it comes up with "cannot execute binary file". Same with ln, mv, mkdir, clear, cp, etc. They all run as root(which makes sense, different files), but I have no idea why this happened. I don't want to stay as root to move around easily. Any idea?

    Read the article

  • Memory cache Ubuntu 9.10 server x86 doesn't work as expected

    - by Matthijs
    We're using an Ubuntu 9.10 server to transfer Ghost-image files. We configured it only with Samba. And the DOS-clients connect to Samba. The latest updates are installed and so far the servers is running fine. When we image 10 pc's with the same image of 2 files of 2GB there's no disk activity. Everything is loaded in the RAM. There's 4GB in the server. But when we use 2 pc's with 2 different image files of 500 MB (8x) files then there's a lot of continuous disk activity. The speed is lower. So it seems that Ubuntu doesn't cache more then one big file. Are there settings to change this behaviour?

    Read the article

  • samba "username map" stopped to work

    - by Kris_R
    It was time to upgrade our group server (new HDs, problems with old installation of DRBD, etc..). Going as usually for CentOS i upgraded whole system from 6.3 to 6.4 The later one came with samba 3.6 as the old one was 3.5. I transferred most of users by copying /etc/password, /etc/shadow and samba accounts with pdbedit. Homes were on nfs-drive. The translation of unix accounts to samba accounts are located in /etc/samba/smbusers. Strangely enough on some windows clients there was problem to connect to samba-shares. In one case the only thing that worked was, instead of giving windows name, to use the unix account. In another one, it was possible to mount network drive and to open it in Windows Explorer, however other applications like "Total commander" at the attempt of opening this drive gave the message "Cannot connect to z:" (sometimes at this moment user/pass were requested). The smb.conf has following entries: [global] security = user passdb backend = tdbsam username map = /etc/samba/smbusers ... [Kris] comment = Kris's Private path = /SMB/Users/Kris writeable = yes read only = no browseable = yes users = krisr printable = no security mask = 0777 force security mode = 0 directory security mask = 0777 force directory security mode = 0 force create mode = 0775 force directory mode = 6775 The smbusers: # Unix_name = SMB_name1 SMB_name2 ... krisr = Kris Of course testparm runs without any errors. I was used from samba 3.5 to outputs of form Mapped user Kris to krisr. Nothing like this happens now. Just message check_sam_security: Couldn't find user Kris in passdb. I read on web that some guys had problem with 3.6 and security = ADS, but these were not helpful for me. I'm seriously thinking about downgrading back to samba 3.5 but before this step I wanted to ask if somebody knows the solution of these problems. p.s. i've asked this question at serverfault but no answer came. Maybe I have more luck with this forum. Sorry for duplicate if any of you reads both.

    Read the article

  • How does SSMS and SQL Server Licensing work?

    - by DrewK
    Could not get a efficient enough answer from MSFT or some of their vendors. Trying to determine exactly how the licensing works before dropping the money on it. Looking to get Server/CAL. We will have the server at our datacenter and then be using SSMS remote on each developers computer. That is, installing SSMS on all developers machine. I am not familiar with MSFT licensing (postgresql & mysql). If I were to pay for the server license and 5 CALs does that mean we can install SSMS locally on each machine. Does each CAL have a specific lic. # that is entered when installing SSMS? We were messing with just the trial edition and the only way I know of installing SSMS is using the full sql server install and choosing only SSMS, it still requires a license number. Any information would be very useful.

    Read the article

  • Git not work on my MacOSX .

    - by koko
    I downloaded and install Git from http://code.google.com/p/git-osx-installer/ . After installing I can't use my Git installation. thomas-macbook-:test zozo$ git init -bash: git: command not found What is wrong? Many thanks in advance.

    Read the article

  • Trouble getting FTP login to work in IIS6

    - by Frank Rosario
    Hello all, I'm trying to setup an FTP site for one of my clients to pickup files from us using IIS6. I've created the FTP site, have set to not isolate users (not necessary as FTP will be read only with authentication). Here's the problem. The FTP is to be password protected, so I turned of anonymous access on the FTP site. I then created a ftpuser account on the machine, and gave it read and browse directory permissions on the ftp's root directory. However, when I go to test the ftpuser login, I get a 530 "ftpuser cannot login" error. However, if I browse to same directory over HTTP (anonymous access turned off as well) and enter the ftpuser login info, I can download files and browse directories successfully. Why is the ftpuser working over HTTP but not FTP? Shouldn't I be able to login over FTP with the ftpuser login information I just created? Thanks in advance, - Frank

    Read the article

< Previous Page | 72 73 74 75 76 77 78 79 80 81 82 83  | Next Page >