Search Results

Search found 82071 results on 3283 pages for 'file url'.

Page 407/3283 | < Previous Page | 403 404 405 406 407 408 409 410 411 412 413 414  | Next Page >

  • Drive system file size

    - by rezx
    When i made a new drive it take some space for system file FAT32 take the less space, then NTFS, then ext4 my question how to know the space will be taken for the system before make the drive, if the drive 1giga or 100giga for FAT32, NTFS, ext4. Edit: when make 10MB drive with FAT32 the size shown 9.9 when make 10MB drive with ext4 the size shown 8.1 the same thing with the bigger size there always some space used and there is no files on the drive, so where this space go, if it for the filesystem how i can calculate the space that will be taken before format the drive

    Read the article

  • Set Permanent System variable through bat file

    - by shyameniw
    I want to change System variables in XP through running a bat file But when i run it i get the error Too many command-line parameters for the following code **set KEY="HKLM\SYSTEM\CurrentControlSet\Control\Sessions Manager\Environment" set PATHxx=%Path% reg add %KEY% /v Pathx /t REG_EXPAND_SZ 5 /d %PATHxx%** How can i fix this??

    Read the article

  • Using relative path in PHP.INI doc_root -- getting No Input File Specified -- running as fastCGI

    - by J.R.
    I'm attempting to run php-cgi under LightTPD on Windows with my doc_root set to one directory up (doc_root = "../Docs") -- however, I get "No input file specified". I've set cgi.fix_pathinfo, and all the other tricks I could find with no success. If I set doc_root to an absolute path, it works fine. How can I make this work? If any additional information is required, I'll gladly provide it. Thanks in advance.

    Read the article

  • Nginx: Rewriting directory path to file

    - by Doug
    I'm a little new to Nginx here so bear with me - I want to rewrite a url like foo.bar.com/newfoo?limit=30 to foo.bar.com/newfoo.php?limit=30. Seems pretty simple to do it something like this rewrite ^([a-z]+)(.*)$ $1.php$2 last; The part that I am confused about is where to put it - I've tried my hand at a some location directives but I'm doing it wrong. Here's my existing virtual host config, where should I implement my rewrite? server { listen 80; listen [::]:80; server_name foo.bar.com; root /home/foo; index index.php index.html index.htm; location / { try_files $uri $uri/ /index.html; } location /doc/ { alias /usr/share/doc/; autoindex on; allow 127.0.0.1; deny all; } location ~ \.php$ { try_files $uri =404; fastcgi_split_path_info ^(.+\.php)(/.+)$; include fastcgi_params; fastcgi_param SCRIPT_FILENAME $document_root$fastcgi_script_name; fastcgi_pass unix:/tmp/php5-fpm.sock; fastcgi_index index.php; } } Thanks!

    Read the article

  • adding response header to Jetty config file

    - by Sam007
    I am trying to add a response header to the Jetty configuration the following command, Access-Control-Allow-Origin: * Ok so this Jetty webserver is the base on which Geoserver is running. I asked this question in the GIS forum but no one was able to answer it so I felt I might get the answer here. I wish to add this response to the header to remove the error Unable to load XMLHttpRequest I wish to know that under which tags should I wrap this response header and where in the file should I save it.

    Read the article

  • DVD (vob file) to online video viewer?

    - by Nick
    I've been sent a DVD which needs to be put onto a website, but I honestly don't even know where to start. Do I simply convert the file using some software to MP4(?!) and then use something like http://videojs.com/ to view it online? I'm really sorry for the vague question, but I want to produce the best quality results, with good compression, good quality and a nice video player interface. Would really appreciate any recommendations. Thank you!

    Read the article

  • How to prevent a file from accessing the internet

    - by Mehper C. Palavuzlar
    There are some exe and dll files belonging to some programs on my laptop which communicate with the internet without my permission (for self-update generally). I want to prevent those files one by one from accessing the internet. I don't want to install any external firewalls. I'm using Windows 7 Home Pro and if this can be done by Windows Firewall, it's OK. I took a look at Windows Firewall but couldn't find a setting to disable file access.

    Read the article

  • Cloned Centos 6.4 websrver for test purpose. Virtual host, .htaccess, redirecting url issue

    - by Shogoot
    I see similar questions, but not my exact challenge. What I have done so far I cloned a prod server over to a vmware to use it as a test server for new functionality I'm going to write. I'm not a sysadmin by trade, but I'm new to this company and I have to do some thing that are outside of my comfort zone (thats a good thing :) ) The prod server has 2 sites on it s1.com and s2.com. In /html/s1/, /html/s2/ there's an .htaccess file under each s*/. Looking like this: RewriteEngine ON RewriteBase / RewriteCond %{QUERY_STRING} id=([0-9]+) RewriteRule ^.* %1.htm RewriteCond %{QUERY_STRING} page=modules/checkout RewriteRule ^.* order.php RewriteCond %{QUERY_STRING} page=pages/sidekart RewriteRule ^.* pages/sidekart.htm The issue is that s1 has a lot of pages that really belongs under a third domain s3, the rule in line 4 and 5 redirects them to /html/s1/. An example of such URL is: s3.com/?page=modules/product&id=521614 I'm trying then to get those URLs (without modifying the URL) to redirect to s3's /html/s3/ server structure, which I set up making a new virtualhost s3 in test servers httpd.conf with a test3.com as servername and changing the other sites to tests1.com and tests2.com, and adding .htaccess also to this s3 root directory, and making a html/s3/ directory structure I populated with an index.html, etc. But, when I take the same URL (s3.com/?page=modules/product&id=521614) changing it to tests3.com/?page=modules/product&id=521614, I get s1's index page showing up in my browser. I've poked around about a day now and i cant figure out why this happens.

    Read the article

  • Sync file system changes in a folder to server on LAN (via SSH or NFS) automatically

    - by Peter Farsinsen
    Hi, I'm looking for a solution that'll automatically sync file system changes in "/folder1" to "/folder2". Any change should be synced (deleted files, new folders etc.) without any confirmation. Preferably the changes will happen to "/folder2" instantly. "/folder1" is on my hdd "/folder2" is on a server on the LAN (can be accessed via SSH or NFS) I'm on OS X 10.6. I can't seem to find just the right tool. Does anyone

    Read the article

  • Chrome: Save as dialogue creates temp file in download directory, I want to change this location

    - by Gabardine
    I've set the download directory for Chrome to be my desktop, but that means that whenever I want to "Save As" it creates a temp file on my desktop until I select the final download destination and close the dialogue. This is deeply frustrating since I browse windowed and I keep seeing the damn things pop up and disappear in the corner of my eye, is there any way to change the directory in which temp files are created in this manner?

    Read the article

  • What are the different file permission codes and what do they mean?

    - by zeckdude
    I am working with a file upload script. I am currently uploading a file and then trying to echo out an anchor linking to that file, but since I used mkdir() with 0700 permissions to upload the file, it won't allow me access to view the file. I am pretty sure the problem I am experiencing is because of the file permission code I used. The problem is I just don't know what all the different file permission codes are and what they mean. Can somebody please list out all the different file permissions and what they each do?

    Read the article

  • Copy and paste RESULTS of a file search

    - by ray023
    In order for me to get the results of a file search in a readable format, I use the following command line in a command prompt: dir *.* /s > myResultList.txt I then open that list in Excel, use fixed width format to get rid of all the stuff I don't want and then I have my list. Seems like a lot to do for something so simple. Does anyone out there have any recommendations for something that would work better than this?

    Read the article

  • How can I instruct nginx to interpret an individual file as php

    - by codecowboy
    A legacy application relies on the following .htaccess rule for an individual file <Files MyFile.html> AddType application/x-httpd-php .html </Files> The new machine has nginx and php-fpm installed with the following config: location ~ \.php$ { try_files $uri =404; fastcgi_split_path_info ^(.+\.php)(/.+)$; fastcgi_pass unix:/var/run/php5-fpm.sock; fastcgi_index index.php; include fastcgi_params; }

    Read the article

  • Run .ipa file on cross platform? [closed]

    - by Fookraa
    Possible Duplicate: iPad Simulator (install ipa on pc) My platforms are Linux and Windows. Now I have a software whose extension is .ipa I am looking for a way to run this file either on Windows or on Linux. There exists Wine which can run .exe files on Linux, similarly is there something which can run .ipa either on linux or windows?? I am also comfortable with getting into virtualization and emulation to get the software work!

    Read the article

  • How to ensure local file is up-to-date or ahead (dropbox sync) before truecrypt auto-mount it?

    - by user620965
    There are a lot tutorials out there that states that dropbox build-in encryption is not secure enought. That tutorials recommands to sync a truecrypt container file to have all files in it securely encrypted. This setup is know to be limited. You can NOT have that truecrypt container file mounted on the same time on more than one location - if you have inserted changes to the contents of the container in more then one location at a time then this setup produces a conflict on the container file in the dropbox system - resulting in one container file for each location. In my case that issue is not relevant - i do not use my data on more than one location at a time. I want to use the auto-mount feature of truecrypt on startup of windows 7 to have a zero configuration environment - and start working right away. But i want to ensure that the local truecrypt container file is up-to-date before truecrypt mounts it automatically - imagine you updated the contents of the container on your primary location and your secondary location was off for a long time. In that case it can take "a long time" till dropbox sync is complete (e.g. depending on your internet connection and the size of the container file). There is a option in truecrypt that ensures that truecrypt do not update the timestamp of the container file - which speeds up the sync, because dropbox client is doing a differential sync then instead of a time consuming full-sync. That is an improvement to that setup, but this do not fix my issue. The question is how to make the auto-mount function wait for the container file to be up-to-date (updated by dropbox)? In contrast: if the file was changed local, but remote file (in the dropbox cloud system) is still old (not jet updated by the sync process / or process is progress), should not make truecrypt to wait for the sync. Suggestions?

    Read the article

< Previous Page | 403 404 405 406 407 408 409 410 411 412 413 414  | Next Page >