Search Results

Search found 13214 results on 529 pages for 'git config'.

Page 263/529 | < Previous Page | 259 260 261 262 263 264 265 266 267 268 269 270  | Next Page >

  • How to use ssh-agent (and graphic passphrase dialog) in Kubuntu 10.10?

    - by halo
    I have recently switched from Ubuntu to kubuntu, both version 10.10. Unfortunately in KDE the ssh passphrase dialog doesn't work out of the box. Everytime my ssh private key is used I need to enter my password. This is neither secure nor comfortable. I have done comprehensive research on the net but only found out dated documentation. Several things I tried didn't work out. Current status: SSH setup working with direct passphrase input ssh-agent running in X session $SSH_AUTH_SOCK set to ssh-agent's socket in X session How to enable ssh-agent for keeping the passphrase in memory for ~15mins and as a bonus always use a graphical dialog for its input? I use asymmetric SSH key pairs for pushing/pulling in Git VCS logging remotely into different server systems

    Read the article

  • Temporarily configuring a default gateway (Windows Vista)

    - by justintime
    I would like to change the default gateway for the wired ethernet port on my laptop- the DHCP server is giving the wrong one. I would like it to revert to the existing config on a reboot. I would prefer a command line way of doing it - so I can easily repeat the command as needed. Do I need ipconfig or route - both of which confuse me.

    Read the article

  • NGINX: dynamic locations stored in DB

    - by chimpanzee
    Is there a possibility to store nginx locations in DB instead of the config to serve them dynamically? The task is to create dynamic URLs for video files based on user's IP and video ID. The idea is when the user visits my website such an dynamic URL is created and added to the db as a new nginx location that exists just for this user and not for others. Or nginx doesn't fit my task and I need to use another tool? Thanks.

    Read the article

  • use Amazon EC2 tools in Capistrano to get the servers to push the code

    - by APZ
    I am trying to use EC2 tools to get all the machines with a particular tag in some type of array in /config/deploy/prod.rb file in Capistrano. Something like this: In prod.rb file: //untested command workers-array[]=$(ec2-describe-instances -F vpc-id=1234 -F tag:Env=prod -F tag:SystemType=worker) for(i=0;i<workers-array.len;i++){ role :worker-A, workers-array[i] } I am not sure how we can do this in capistrano, am new to ruby too. Guys any help on this would be really appreciated.

    Read the article

  • How to configure apache's mod_proxy_html to work as an ajax proxy?

    - by dcerecedo
    I'm trying to build a web site that let's you view and manipulate data from any page in any other website. To do that, I have to bypass 'Allow Origin' problems: i'm loading the other domain's content in an iframe and i have to manipulate its content with javascript downloaded from my domain. My first attempt was to write a simple proxy myself, requesting the other domains page through a server proxy coded in Java that not only serves the content but rebuilds links (src's and href's) in the content so that the content referenced by these links alse get downloaded through my handmade proxy. The result is not bad but has problems with url's in css and scripts. It's then that i realized that mod_proxy_html is supposed to do exactly all this job. The problem is that i cannot figure out how to make it work as expected. Let's suppose my server runs in my-domain.com and to proxy and transform content from another domain i'd make a request like this: my-domain.com/proxy?url=http://another-domain.com/some/content I'd want mod_proxy_html to serve the content and rewrite following URLs in http://another-domain.com/some/content in the following ways: Absolute URLs not from another-domain.com: no rewritting Relative from root urls:/other/content - /proxy?url=http://another-domain.com/other/content Relative urls: other/content - /proxy?url=http://another-domain.com/some/content/other/content Relative to parent urls: ../other/content - /proxy?url=http://another-domain.com/some/other/content The url should be specified at runtime, not configuration time. Can this be achieved with mod_proxy_html? Could anyone provide a simple working configuration to start with? EDIT 1-First approach The following site config will work fine with sites that use absolute url's everywhere like http://www.huffingtonpost.es/. Youc could try on this config on localhost: http://localhost/asset/http://www.huffingtonpost.es/ <VirtualHost *:80> ServerName localhost LogLevel debug ProxyRequests off RewriteEngine On RewriteRule ^/asset/(.*) $1 [P] ProxyHTMLURLMap $1 /asset/ <Location /asset/> ProxyPassReverse / ProxyHTMLURLMap / /asset/ </Location> </VirtualHost> But as explained in the documentation, if I hit a site using relative url's, I'd like to have these rewritten on the html via mod_proxy_html. So I shoud change the Location block as follows: <Location /asset/> ProxyPassReverse / #Depending on your system use one line or the other #Ubuntu: #SetOutputFilter proxy-html #any other system: ProxyHTMLEnable On ProxyHTMLURLMap / /asset/ </Location> ...which doesn't seem to work. Comments, hints and ideas welcome!

    Read the article

  • Focus previous window when I close one instead of one of the same application in Mac OS X?

    - by lilydjwg
    The situation is: I have one MacVim open for coding. Then I switch to iTerm to run git commit, which will open a new MacVim for editing the commit message per my setting. The problem is, when I finish writing the commit message and close the MacVim window, the previous iTerm window is not focused. The other MacVim window is focused instead. Is there any way I can change this behaviour? I just want the previous window get focused when I close a window.

    Read the article

  • How can I make .vimrc read from an external file?

    - by GorillaSandwich
    I'd like to modify my .vimrc to read the value of a variable from an external file. How can I do this? Specifically, a friend and I share a git repo with our .vim files, but there are a few small differences in what we want in our configs. So most of the file is common, but we use if statements to determine whether to load user-specific sections, like this: let whoami = "user2" if whoami == "user1" ... After checking our common .vimrc out of source control, we each have to change the let whoami assignment so our own section will be loaded. Instead, I'd like to keep a separate file, which can be different for each of us, and from which vim will load that variable value. Maybe another angle on this is: Will vim automatically read all the files in my .vim directory? If so, we could each put a symlink in there called username.vim, and link that to an external file that would be different for each of us.

    Read the article

  • How do I disable Aero transparency in Firefox 4?

    - by Titus
    I noticed that the transparent interface in Firefox 4 for Windows 7 is really curbing performance. After Googling for a method to disable Aero for Firefox, however, I got no results. There just doesn't seem to be a straight forward (such as a simple Boolean to turn it off in about:config) method to get rid of Aero effects. Is it even possible in the Firefox 4 beta? I'm using Firefox 4 beta 12 (soon to be beta 13 after the update).

    Read the article

  • safely remove webserver from nginx backent pool?

    - by Dean Hiller
    We have 4 webservers behind nginx being hit with 262 events/second. I would like to tell nginx to stop sending requests to that server. If I remove the server from the file and reload the file, aren't all requests in process dropped on the floor then as nginx no longer knows that server. What can I add to the config so it slowly drops that server out of the pool and I can wait for current requests to complete.

    Read the article

  • Creating a fallback error page for nginx when root directory does not exist

    - by Ruirize
    I have set up an any-domain config on my nginx server - to reduce the amount of work needed when I open a new site/domain. This config allows me to simply create a folder in /usr/share/nginx/sites/ with the name of the domain/subdomain and then it just works.™ server { # Catch all domains starting with only "www." and boot them to non "www." domain. listen 80; server_name ~^www\.(.*)$; return 301 $scheme://$1$request_uri; } server { # Catch all domains that do not start with "www." listen 80; server_name ~^(?!www\.).+; client_max_body_size 20M; # Send all requests to the appropriate host root /usr/share/nginx/sites/$host; index index.html index.htm index.php; location / { try_files $uri $uri/ =404; } recursive_error_pages on; error_page 400 /errorpages/error.php?e=400&u=$uri&h=$host&s=$scheme; error_page 401 /errorpages/error.php?e=401&u=$uri&h=$host&s=$scheme; error_page 403 /errorpages/error.php?e=403&u=$uri&h=$host&s=$scheme; error_page 404 /errorpages/error.php?e=404&u=$uri&h=$host&s=$scheme; error_page 418 /errorpages/error.php?e=418&u=$uri&h=$host&s=$scheme; error_page 500 /errorpages/error.php?e=500&u=$uri&h=$host&s=$scheme; error_page 501 /errorpages/error.php?e=501&u=$uri&h=$host&s=$scheme; error_page 503 /errorpages/error.php?e=503&u=$uri&h=$host&s=$scheme; error_page 504 /errorpages/error.php?e=504&u=$uri&h=$host&s=$scheme; location ~ \.(php|html) { include /etc/nginx/fastcgi_params; fastcgi_pass 127.0.0.1:9000; fastcgi_intercept_errors on; } } However there is one issue that I'd like to resolve, and that is when a domain that doesn't have a folder in the sites directory, nginx throws an internal 500 error page because it cannot redirect to /errorpages/error.php as it doesn't exist. How can I create a fallback error page that will catch these failed requests?

    Read the article

  • What are the differences between the "generic" and "server" kernel images provided by Ubuntu?

    - by dcrosta
    In particular, I'm wondering if there are any patches or config adjustments made to the disk cache size in the server edition. I'm running on a small system (256M RAM), and would like to experiment with keeping the disk cache size smaller so that there's more memory available for applications. I've found this page at Ubuntu's website, which neither answers my questions nor is about the 9.04 release.

    Read the article

  • How can I trick ssh to connect using different configurations based on current location?

    - by Sorin Sbarnea
    Here is the following use case: wanting to ssh server1 in 3 scenarios: server1 is on current network, go for it server1 is not on current network and you need to use a gateway (gw1) to access it, not a big problem as you can use ~/.ssh/config to define tunnelling for this you are on a network that does not allow you to access the first gw1, so you'll use gw2 for that. Desired outcome: be able to use the same command and be able to connect to the server.

    Read the article

  • Advice on Auditing VPN Connections

    - by WhiteAdam
    I've inherited a network that is truly a mess and the VPN config is amazingly long with confusing group names and very little documentation. I am curious to see how others have gone about auditing these types of connections. Are there any forms out there? What information should I collect? I assume things like Customer, Department, Technical Lead, and etc. Thanks in advance for any advice on this nightmare.

    Read the article

  • Can't connect to Gmail server via Mail.app in Mac OS X 10.6.3

    - by Johnny
    I've added my gmail account to Mail.app It worked find in previous days, and downloaded thousands of previous mails. But now, it can't connect to gmail server for days. What's the matter here? Here is my config in account setting: Account Type: Gmail IMAP Email Address: [email protected] Incoming Mail Server: imap.gmail.com User Name: [email protected] Password: xxxxxx And also, is there any means that I can view the transaction log of Mail.app? Maybe there I can find more information.

    Read the article

  • Custom settings with PHP on Nginx

    - by miki
    I have multiple websites setup using Nginx and Apache, but when I try to add a vhost based PHP directive using fastcgi_param PHP_VALUE , the value get added to all of my vhosts. Checking the value of the PHP parameter from PHP-CLI remain the actual one for whole server. e.g. I used fastcgi_param PHP_VALUE "memory_limit=512M" for a domain based nginx config, but it will be propagated to all domains on server. The 'php -i | grep memory_limit' show 128M though Not sure what I am missing

    Read the article

  • Monit Not Starting

    - by heavymark
    Installed Monit per http://wiki.mediatemple.net/w/%28dv%29_4.0_-_Making_It_Better_::_Installing_Monit monit -t shows syntax is OK. My config file is: http://pastie.org/5031846 Running: /etc/init.d/monit start Results in: Starting monit: It's suppose to say: Starting monit: Starting monit daemon with http interface at [*:2812] [ OK ] This is on Centos 5.8. Is there anyway I can test why it's not starting properly or how to resolve it?

    Read the article

  • Authenticated proxy in Windows command prompt

    - by tlvince
    My Internet access is delivered through an authenticated proxy. This is setup correctly in the "Internet Options" control panel applet and all browsing/GUI apps work fine. For cmd, proxy settings are typically set by exporting http_proxy. The applications I'm trying to use are git and Putty's pscp.exe. I've tried the following syntax to no avail: set http_proxy=http://username:[email protected]:port/ I've also experimented with Netsh.exe and ProxyCfg.exe with no luck. How does one configure an authenticated proxy in the Windows command prompt?

    Read the article

  • Application deployment with Puppet

    - by michaeltwofish
    I'm new to Puppet and while I've been using *nix systems for many years, I've never worked as a sysadmin or in ops. I'm currently writing Puppet manifests for hosting a set of (PHP/MySQL/MongoDB, code in git) web applications. Clearly Puppet needs to have some knowledge of the actual applications because I'll set up a virtual host for each one, but I'm not sure whether Puppet should be managing things like code deployment and database creation. Is Puppet an appropriate tool for application deployment? If not, can you recommend a more appropriate tool?

    Read the article

  • change socket to other then default in phpPgAdmin

    - by DanFromGermany
    I need to change the socket phpPgAdmin connects to in its config. // Hostname or IP address for server. Use '' for UNIX domain socket. // use 'localhost' for TCP/IP connection on this computer $conf['servers'][0]['host'] = '/opt/jasperreports-server-cp-5.1.0/postgresql/.s.PGSQL.5432'; this does not work (even without the last part .s.PGSQL.5432). The path is correct, because I can connect through: :~# psql --host=/opt/jasperreports-server-cp-5.1.0/postgresql/

    Read the article

  • raid 0 failure, drives look fine

    - by Alan
    Hello, after a lovely blue screen my vista 64 machine decided to reconfigure one of my drives to no longer be part of my raid volume. So now my raid fails as it only has one member disk. This happened to me about 6 months ago and I just changed the disk in question back to a raid disk and all was well. However I cant seem to find that option in my bios or raid config anymore :( Any help would be appreciated

    Read the article

  • How do I use rsync to sync my repo on my mac to university server where I have ssh access?

    - by snihalani
    I need to use my university's ssh access and run my programs there for testing. I don't have sudo access there. It doesn't have vncserver there either. I would work with vim and make but I need git at least. Now I am looking into rsync to sync my current source directory into a remote directory and I'll ssh into the directory and run my make file to test it. I am looking at the man page of rsync and it looks very complicated. Can anyone please help me with this? I have googled in superuser and all commands seem different for different cases. Can anyone please help me with this?

    Read the article

  • Connection closed by remote host followed by Connection refused

    - by Khosrow
    All of a sudden my ssh connection to server has been damaged. Here is what's happened: $ ssh -vvv -p <PORT> -l <USER> <HOST> OpenSSH_5.3p1 Debian-3ubuntu7, OpenSSL 0.9.8k 25 Mar 2009 debug1: Reading configuration data /home/khosrow/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Applying options for * debug2: ssh_connect: needpriv 0 debug1: Connecting to <HOST> [<IP>] port <PORT>. debug1: Connection established. debug1: identity file /home/khosrow/.ssh/identity type -1 debug1: identity file /home/khosrow/.ssh/id_rsa type -1 debug1: identity file /home/khosrow/.ssh/id_dsa type -1 ssh_exchange_identification: Connection closed by remote host I've recently updated the box with yum update and sshd got updated as well. I honestly don't know if this caused any damages or not. But it's prompted that /etc/ssh/sshd_config was stored as /etc/ssh/sshd_config.rpmnew which was quite normal. I've seen similar posts while googling, but almost all of them suggests that I should check /etc/hosts.allow and /etc/hosts.deny, which in my case, I can't. I can not connect to the box to see what's going on there. I rebooted the box, through web interface of server provider, and it even got worse. I'm now getting this: $ ssh -vvv -p <PORT> -l <USER> <HOST> OpenSSH_5.3p1 Debian-3ubuntu7, OpenSSL 0.9.8k 25 Mar 2009 debug1: Reading configuration data /home/khosrow/.ssh/config debug1: Reading configuration data /etc/ssh/ssh_config debug1: Applying options for * debug2: ssh_connect: needpriv 0 debug1: Connecting to <HOST> [<IP>] <PORT>. debug1: connect to address <IP> port <PORT>: Connection refused ssh: connect to host <HOST> port <PORT>: Connection refused with both <CUSTOM_PORT> and default 22 ports. I would really appreciate if anyone could help me on this.

    Read the article

  • Where does Chrome store its bookmarks in Ubuntu 11.10?

    - by Alan Wood
    I looked at all the other posts on this but can't find the directories mentioned (~/.config/google-chrome/Default/Bookmarks, it's a JSON file.). Being a 2 day Newbie to Ubuntu/Linux I would like to know if the location has changed in the latest version or if not how I locate the directory indicated. I have logged in as root and searched for the folder and can't find it although I imported my bookmarks from a html file so I know that they must be saved somewhere.

    Read the article

  • In-House DropBox

    - by beardedlinuxgeek
    Dropbox is perfect, but as a company, no one can host anything worthwhile on servers that we don't control. So I've been tasked with coming up with a Dropbox alternative, something in house. GlusterFS is nice, but no offline access. SparkleShare uses Git which isn't great for large files. It also doesn't have windows ports. Any other options? If I were to roll out my own from scratch, what do you think the based way to go about doing this would be?

    Read the article

< Previous Page | 259 260 261 262 263 264 265 266 267 268 269 270  | Next Page >