Search Results

Search found 11453 results on 459 pages for 'apache axis'.

Page 376/459 | < Previous Page | 372 373 374 375 376 377 378 379 380 381 382 383  | Next Page >

  • I installed XAMPP on Ubuntu, and its running without any of its files.

    - by CDeanMartin
    I installed XAMPP on Ubuntu, and its running without any of its files. Well, not really, but it sure seems that way. I followed the directions on the Apache Friends website to the letter. XAMPP works fine, it's sample apps run like a charm. But I have no idea how it is running. The installation should have created an otf folder, or a XAMPP folder, or an htdoc file according to the tutorials. But there is no trace of any of these files or folders anywhere in the file browser. Are there any linux utilities as good as the Windows Explorer that can find my missing files? I particularly need to find the var/www folder to put my .php files in.

    Read the article

  • Running nph-script.cgi keeps outputting Server details at the end

    - by wgewweg
    I am running a nph-script.cgi on my server. The server keeps adding HTTP/1.1 200 OK Date: Thu, 05 Nov 2009 02:28:53 GMT Server: Apache/2.2.8 (Ubuntu) PHP/5.2.8-1hardy~ppa1 with Suhosin-Patch mod_perl/2.0.3 Perl/v5.8.8 Content-Length: 0 Keep-Alive: timeout=15, max=100 Connection: Keep-Alive Content-Type: text/plain X-Pad: avoid browser bug At the bottom of each page loaded via the .cgi script. why is this the case? How do I remove this annoying message that is appended to all pages ?

    Read the article

  • Editing remotely the PHP files on a Centos server

    - by Alex2012
    I have a intranet web server (Centos 6, Apache, PHP) to which I would like to give access to a developer. He will connect by remote desktop from Windows 7 to Ubuntu 12.4 and from here by SSH to /var/www/html folder where it has to create and edit the files. This solution was chosen because: - I could not make a remote desktop connection from Windows to Centos - The web developer need some editor for PHP files and is not allowed to install software on Windows 7 machine - it is more a test solution ( we are all learning to use Linux). When the developer is connected from Ubuntu to Centos by SSH (SFTP) he could save the changes only if on Centos the account used to connect has ownership to that folder. Can you please tell how can I give all required rights. I tried different solutions found on Internet but without to much success. Are there other way to connect to Centos server?

    Read the article

  • Asking for brief explanation of reverse proxying and recommended software

    - by 80skeys
    I need to set up a reverse proxy where the backend web server is serving https pages only. I've never set up a reverse proxy and would like a brief overview of how it works. One of my questions is whether the proxy needs to run https also, or does simple http suffice? Second question would be whether to use Apache, varnish, nginx, or squid. This is for an internal site for a small company, so not a lot of traffic expected. Maybe a few dozen users each day.

    Read the article

  • Restarting Nagios Using PHP

    - by X-Ware
    I am making a tool that is interacting with NAGIOS where some config files should be added so a restart will be needed. What I need to know is how to restart NAGIOS using PHP code since this tool is written in PHP .. when I try to do this using: shell_exec("service nagios restart"); changes do not take place but when I do this manually by the console all changes I did using the PHP script are applied ... after 2 minutes research I found that I am asking linux to execut this command while I am logged in as apache user so I changed the command to: shell_exec('echo "mypass" | sudo -S service nagios restart'); still having the same problem ... new config files are not read until I restart manually any suggestions will be appreciated :)

    Read the article

  • Upgrade PHP 5.1.6 to 5.2.x on Fedora 6 - Solutions?

    - by nicorellius
    I have a CRM on an old Fedora server, running Apache 2.2.6, MySQL 5.0.27, and PHP 5.1.6. I see a couple of options to fix my problem, but I thought I would consult the pros on this one here at Server Fault to see if I'm missing anything. This is after spending a good while searching the net for solutions. Option 1 Upgrade Fedora, step-wise, to Fedora 7 and then to 8, and then the PHP packages I need would be available. I know this is hazardous, because of repository location, dependencies, etc. Option 2 Upgrade using this method. This may or may not work, and so would be considered a crap shoot. I consulted some IT people and they say the best best is to back up data and build a new server with updated hard and software. I happen to agree, but wanted to exhaust all my resources first.

    Read the article

  • FTP Server with MySQL access, and POST notification

    - by TIW
    Im looking for an FTP server solution, that we can host either internally on a dedicated server, or on Rackspace Cloud/AWS, that provides a HTTP POST notification when a file is uploaded, and allows user accounts to be created either through an API or MySQL database. There are several offerings that provide email notification - but has anyone come across anything that matches the above requirements. BrickFTP being a IaaS system is an option, but we would prefer something hosted in house. I don't believe the standard FTP servers provided with Apache can do the above ... can they?

    Read the article

  • svnserve, Mac OS X Lion Server and user accounts

    - by Casper
    I am trying to use the "built-in" user accounts of my Mac OS X Lion server to grant access to repositories. Currently I am using the conf/passwd file (which works), but I don't like the fact that the passwords are visible as plain text. Is there a way to "connect" the svnserve access restriction to the normal user accounts that are already on the system? Thanks PS: I am not wanting to connect via Apache/WebDav - there I know it works. I want to connect via svn:// using the existing user accounts.

    Read the article

  • Can't connect to localhost via browser. Can ping localhost.

    - by Sceptre
    I'm trying to connect to localhost through my browser to learn some apache tomcat stuff. When I tried to connect to localhost through Firefox, I couldn't; when I tried through IE, I could the first time, but not after that. I'm using Windows 7, and changed the hosts file to point localhost to 127.0.0.1. I can successfully ping localhost and 127.0.0.1. I have tried turning off my antivirus and my Windows Firewall, but to no avail. What am I doing wrong?

    Read the article

  • htaccess with wildcard SSL

    - by Ericko
    We have a Wildcard SSL Certificate that is supposed to work on any subdomain of a given domain. So in this server we have this file structure: /home/DOMAIN/public_html/subdomainx /home/DOMAIN/public_html/subdomainy etc... Now, the Certificate is installed, but when you visit any subdomain over https (example: hxxps://subdomainx.domain.com ) it points to /home/DOMAIN/public_html/index.php We need that when you visit a subdomain via https hxxps://subdomainx.domain.com That it points to the the same directory that it's http equivalent: /home/DOMAIN/public_html/subdomainx Our provider tells us that this is not possible, that the current behaviour is correct, and that to achieve this we need to do it with htaccess. I've tried a few things, incluiding this solution, that seems to be what I need: http://stackoverflow.com/questions/5365612/advice-on-configuring-htaccess-file-to-redirect-http-subdomain-to-https-equival But can't get it to work. Any tips? Thanks. Added: The server is Apache.

    Read the article

  • HAproxy to web host sub directory?

    - by daemonza
    Hi for reasons outside my control, I need to load balance two servers, that run a non-virtual host enabled app on IIS. Normally in HAProxy I would load balance servers(apache, tomcat, etc) like this : acl is_www_example_com hdr_end(host) -i www.example.com use_backend www_example_com if is_www_example_com backend www_example_com balance roundrobin cookie SERVERID insert nocache indirect option httpchk HEAD / HTTP/1.0 option httpclose option forwardfor server node1 192.168.1.1:80 cookie node1 server node1 192.168.1.2:80 cookie node1 Which will route to the node 1 and node 2 server and serve up the virtual host site. if I need to route to www.example.com/application/data How would I be able to do it, with the above example, if at all even possible?

    Read the article

  • solutions for a webserver dedicated to manage permissions/ACL and (reverse) proxying API servers?

    - by giohappy
    I'm considering various layouts to expose various HTTP API services (running on their own differents servers) through a frontend server dedicated to manage permissions on behalf of the API services. I've considered various options, from the classical ones like Nginx, Apache, etc. to HAProxy, passing by the various Python webserver solutions like Tornado, Twisted (which gives me the opportunity to implement my own ACL system easily). The foundamental feature is high performance and scalability, and the ability to manage fine grained ACL rules (similar to the HAProxy ACL system) I would like to know what is a suggested approach to setup what I need, and if (opne source) ready-to-use solutions are already available dedicated to this.

    Read the article

  • HTTPD Restrict Acces to File via Web App

    - by ssgriffonuser
    I would like to restrict access to a file until the user has paid for it. This is easy enough if I served the file via a web app, but it seems more efficient to have httpd serve the file. Is there a way to get apache to ask a web app if it can serve a file before serving it? As I'm typing this it seems like it should be trivial but I'm not coming up with a good answer via google. Thanks in advance for any help.

    Read the article

  • How do shared hosting servers keep executing code from crossing accounts?

    - by acidzombie24
    I am kind of curious, how does a hosting server support multiple users with php but keep each user away from the other code? The 'easy' solution i thought were file permissions. So every user can have www-data belong to their group and the server would have executing access but the users cant access the others file. But then i realize the user running the php would be www-data who has permission to read everyones data. So how does a shared host prevent this from happening? PS: I personally use nginx (with fastcgi php). But i am somewhat familiar on how apache works.

    Read the article

  • php.ini settings change not taking effect for large file uploads

    - by user51347
    My server was just reprovisioned, and my application which uploads large (100M+) files now breaks upon re-installation. the symptom is quite consistent : smaller files (8mb in my tests) upload just fine. Larger files cut off at quite close to the same point every time given a particular computer. a file that fails at 26% will fail at just about the same every time. One that takes 1:40s to fail will take within 2 seconds of that every time, before failure. I have set my php.ini settings extravagently : post_max_size = 512M upload_max_filesize = 512M max_input_time = 3600 max_execution_time 3600 Is there possibly a setting at the Apache Level which would override PHP?

    Read the article

  • Disabling the shell of user "daemon" (/bin/false)

    - by BurninLeo
    on a Linux system there are lot's of users by default: daemon, bin, sys, games, etc. According to my /etc/passwd most of these users have a shell assigned (/bin/sh) which seems some kind of insecure to me. My naive thinking would say: Only give those users a shell that may login to the server. Is my thinking wrong? If not completely wrong: Can I disable the shell for "daemon" and "www-data" without having side effects (e.g. the system wont start or the Apache PHP cannot excute system calls)? Thanks for your hints!

    Read the article

  • PHP5.3 FastCGI doesn't use global config's values

    - by mega.venik
    There's a Centos6.3 system. Apache 2.2.15 + mod_fcgid + PHP 5.3.3 There's a problem with date.timezone value. It's mentioned in the global /etc/php.ini like this: date.timezone = "Europe/Moscow" And doesn't mentioned in user's local php.ini. As a result, I'm getting lot's of warnings like: Warning: date() [function.date]: It is not safe to rely on the system's timezone settings. You are *required* to use the date.timezone setting or the date_default_timezone_set() function. In case you used any of those methods and you are still getting this warning, you most likely misspelled the timezone identifier. We selected 'Europe/Helsinki' for 'EEST/3.0/DST' instead in ... Including the date.timezone parameter into the user's php.ini solves the problem, but I don't think, that it's the best solution. Maybe someone have faced this problem and can give an advice? Thanks! P.S. Creating /etc/php.d/timezone.ini with the timezone info aslo does nothing:(

    Read the article

  • Rewrite 2 different directories with htaccess?

    - by jason
    I have a tricky problem (for me at least). I'm trying to rewrite / to a folder /webroot/www. I have some simple code and it works: RewriteRule ^$ /webroot/www/ [L] However at the same time if the URL starts with components, followed by anything else (ex. foo, as in /components/foo), and foo is an actual directory that exists inside components, I should rewrite to /components/foo/www instead. How can I achieve that? I can't seem to figure it out. I'm using Apache with .htaccess.

    Read the article

  • Point IP to site for testing before delegation?

    - by cosmicbdog
    I'm migrating to a new server and would like to test before redelegating the domain over. I have some familiarity with setting up apache virtual hosts, but limited knowledge. How can I go about setting this up? My server already has a domain delegated to the server, and the IP by default has been setup to point to that. I've been told I can just add an entry to /etc/hosts/ which I haven't been able to understand what I can add in there to make this work. Any pointers would be greatly appreciated!

    Read the article

  • phpMyAdmin tried to connect to the MySQL server, and the server rejected the connection [closed]

    - by Sagar Thapaliya
    i have some problem i had skpe on my laptop and unfortunately skype used 80 port for network and wamp server 2.0 Apache Version : 2.2.10 PHP Version : 5.2.7 i changed the port of wamp server through Listen 12.34.56.78:80 Listen 80 Listen 12.34.56.78:80 Listen 8080 and i able to do somehow but i can't access the phpmyadmin please help me i hope i can get my solution from all of you soon my errors seems this i have mentioned downstairs below phpMyAdmin tried to connect to the MySQL server, and the server rejected the connection. You should check the host, username and password in your configuration and make sure that they correspond to the information given by the administrator of the MySQL server.

    Read the article

  • Cannot Access Server from External IP (Router)

    - by mindoftea
    We have an Ubuntu 10.04 LAMP server running on site (Apache is on port 80). It is running fine through the LAN, but I cannot access it using its external IP address through the Netopia 3000 Series modem by which it is connected to the internet. I have tried using what Netopia calls "Pinholes" (Port Forwarding) to make the server accessible, but a telnet to the external address just gives me "connect to address x.x.x.x: Operation timed out." I have also tried enabling "Services" and "IP Passthrough" on the router, but it gives the same result as above. How can I enable the server to be accessed through its external IP address? Because it connects fine locally, it would seem to be a problem with the Netopia router. Update: Booting the server in GUI mode instead of text mode solved the problem instantly. Any ideas why?

    Read the article

  • Can't access link in network using fully qualified domain name

    - by user1033715
    I have installed windows server 2003 and configured Domain controller (domain name - xyz.com) and DNS service. for that I have configured fully qualified domain name as server.xyz.com also I have installed apache tomcat with port 8080 on that server and accessed link successfully using "http://localhost:8080", "http://ip address of server:8080", "http://server.xyz.com:8080". but its working for local machine, and when I tried to access it from another machine in same network using "http://ip address of server:8080" its worked for me. but when I tried it using fully qualified domain name i.e. "http://server.xyz.com:8080" it's giving me error, "Could not connect to server.xyz.com" Please guide me getting this setup done. I need to be able to access this link "http://ip address of server:8080" as "http://server.xyz.com:8080" outside my network Any suggestion are highly appreciated..

    Read the article

  • Cannot log-in in the install of phpmyadmin on Ubuntu

    - by Sylario
    I just installed phpmyadmin and i am trying to login in /phpmyadminpath/setup. I am greeted by an invite similar to an apache protected folder (log/password displayed in a browser pop-up, not an html popup) asking for login pass. I cannot login using system password(root) or mysql passsword(also root) I already have rails app using this mysql, and i checked the mysql password in the config files. I cannot run any mysql command in my console because of the following error : ERROR 1045 (28000): Access denied for user 'root'@'localhost' (using password: NO) What is wrong with phpmyadmin and mysql?

    Read the article

  • phpmyadmin throws a blank page after installing Zend CE server

    - by infotoknowledge
    I installed LAMP on Ubuntu 12.0.4 LTE. Then I installed phpmyadmin. Everything was good. I used phpadmin to do lots of DB administration.... Then I installed Zend CE server. Now, phpmyadmin throws a blank page. No login screen. So, I reinstalled my VPS in the same sequence again. No luck. Same behavior. This time, I installed Apache, Mysql, PHP and Zend server in that order. Then I installed phpmyadmin....Unfortunately, still blank page on phpmyadmin. Any help?...Thanks in advance.

    Read the article

  • Path erased in Debian

    - by Lyon83
    I'm trying to deploy a rails app in Debian, using Apache/Passenger. I was trying to fox a problem with some GEMs and in the process I put executed this in console: export PATH=/var/lib/gems/1.8/bin/:${vendor/cache} Now my path environmental variable is gone, or at least its content. My server is running under Debian 6. Is there a way to recover my path info? Or at least can someone point me where to find the file where that variable i s stored? Some help please. This is a BIG problem for me. Thanks in advance!

    Read the article

< Previous Page | 372 373 374 375 376 377 378 379 380 381 382 383  | Next Page >