Search Results

Search found 37539 results on 1502 pages for 'google friend connect'.

Page 291/1502 | < Previous Page | 287 288 289 290 291 292 293 294 295 296 297 298  | Next Page >

  • Google revoit ses capacités de stockage Cloud à la baisse et ses prix à la hausse, presque tous les utilisateurs de Drive concernés

    Cloud : Google revoit ses espaces de stockage à la baisse Et ses prix à la hausse, du pain bénit pour la concurrence ? Encore plus de « nettoyage de printemps » est à prévoir pour ses services en ligne, prévient Google. Il s'agit en fait d'un toilettage d'automne. Et autant le dire tout de suite, il n'est pas porteur de bonnes nouvelles pour les utilisateurs de Google Drive et de Picassa. Google annonce en effet que les capacités ces deux espaces de stockages « seront consolidées » et que la capacité totale gratuite sera désormais de 5 GB « pour les deux services ». Autrement dit, l'espace gratuit total perd 1 GB (- 16.5%) puisque auparavant Google Drive proposait déjà ces 5 ...

    Read the article

  • App Engine charges in November 2011 [closed]

    - by broiyan
    I had a billing enabled test application on Google App Engine left over from early 2011. I have not received a bill in many months because I have not been hitting the URL and according to the activity monitor, nobody has. Then unexpectedly in November 2011, I received 2 bills in as many weeks for quite minimal amounts. Checking the monitor it looks like nobody has been hitting the URL and according to the SQL-like search, there is nothing in the Datastore. I know that GAE has left the "preview" in recent weeks but I am not sure how that would affect what is essentially a dormant application with no Datastore objects. Has GAE started charging for completely unused applications in recent weeks? Edit: Most of my applications were already disabled and I have just disabled the only one that was enabled but unused the past several months. If I get another bill next week that should be informative.

    Read the article

  • Sudden drop of pageviews/visit and increase of bounce rate in Analytics

    - by Tebb
    Google analytics stats: 04 june 2012 Visits: 4.423 Unique visitors: 3.558 Pageviews: 77.352 Pageviews / visit: 17,49 Visit length: 00:06:26 Bounce rate: 1,09% 05 june 2012 Visits: 4.652 Unique visitors: 3.825 Pageviews: 45.087 Pageviews / visit: 9,69 Visit length: 00:06:45 Bounce rate: 19,60% From one day to another the bounce rate went from 1% to 19%, the pageviews dropped by half so did the pageviews/visit. The only thing I changed (If I remember correctly) on the site, was an advertisment that used a javascript. Could this be the reason? and, if it is, how can I know which one is the real stats?

    Read the article

  • How to smartly optimize ads on website

    - by YardenST
    I've a content website that presents ads. Now, my team want to optimize it for a better experience for the users. (we really believe our ads are good for our users.) We are sure that every website deals with this issue and there must be some known ways and methods to deal with it, that smart people thought of before. so what i'm looking is a tested, working method to optimize ads. for example: if i was asking about optimizing my website in Google, I would expect you to answer me: learn SEO if i was asking about optimizing the use of my website: usability testing. navigation: information architecture what is the field that deals with optimizing ads?

    Read the article

  • Can we 301 redirect to a new page, but still publish the old content somewhere else?

    - by KBS
    We have a page on the site which ranks well for an SEO term (top 5) but contains old information. We have added a new page but Google doesn't rank it that well. Information on these pages is time sensitive. Old: example.com/2013-related-information.html New: example.com/2014-related-information.html Obvious solution is to delete old page and do a 301 redirect to the new page. Now, can we still keep the old page by giving it a new URL. example.com/2013-related-information.html is redirected to example.com/2014-related-information.html example.com/2014-related-information.html is recreated with a new address such as example.com/new-2013-related-information.html What we are trying to do is to send the user to the fresh page but still not destroying the record copy if someone wants to go and dig up the old information.

    Read the article

  • Why is Cloudflare waiting for name servers for over 4 days?

    - by user29175
    I've registered for Cloudflare's free plan and have completed the process of redirecting the DNS as instructed, including changing the name servers. This was done 4 days ago. The problem is that cloudflare is giving me: "websites" -- "Finishing up. Waiting for your name servers to change to * Please allow up to 24 hours to complete this process (info)" "dashboards" -- "Analytics data could not be loaded. You do not have any initialized zones" I can see via traceroute that CloudFlare is the DNS to my site. Also, somehow this has messed up with my google analytics account so I have no idea if I get visitors to my site or not. What should be done to fix this?

    Read the article

  • Getting a lot of '/_' errors from webmaster tools

    - by Vermino
    I'm using a WordPress site and I thought I got all the kinks out of it. For some reason Webmaster Tools is crawling my website and showing a lot of 404 errors which are from /_ like additional pages that I've never created. I just can't figure out what is creating these for Google crawlers and then displaying a 404. My robots.txt is here. My sitemap (created by the Yoast plugin) is here. I have Yoast and Jetpack plugins installed. What could be causing these links to appear

    Read the article

  • Why does the homepage rank in exact match and not in broad anymore? [closed]

    - by Claude Pelanne
    My website was ranking on page one for "walking shoes reviews" in broad and exact. It has disappeared in the rankings for broad and yet still ranks in exact match. No indication in the Webmaster tools of anything wrong yet Google analytics shows no more activity. his occurred sometime after Panda and Penguin. Not sure which because I didn't check closely for a while, this is a fun site for me. I walk for the enjoyment and health benefits so I built this site.

    Read the article

  • Installation of ANYTHING failed

    - by Nervosa
    I got an issue concerning Chrome. It launches now perfectly - still trying to install something else i see - You might want to run 'apt-get -f install' to correct these: The following packages have unmet dependencies: google-chrome-stable : Depends: lib32gcc1 (>= 1:4.1.1) but it is not installable Depends: lib32stdc++6 (>= 4.6) but it is not installable Depends: libc6-i386 (>= 2.11) but it is not installable E: Unmet dependencies. Try 'apt-get -f install' with no packages (or specify a solution). I remember i deleted some folder concerning chrome when it wasn't launchable - don't actually remember what directory exactly. Now - when i try this - "sudo apt-get -f update && sudo apt-get -f install" - i come across an error: sh: 0: getcwd() failed: No such file or directory Seems that deletion was fatal. Got any ideas? Thanks.

    Read the article

  • Help with tracking sub domain

    - by roobus
    I currently have my app's marketing/external website on the root level, e.g. http://example.com My web app itself is hosted at: http://app.example.com What's the best strategy to set-up Google Analytics tracking for both of them? Should I create a separate web property? Also, what's the difference between creating a new web property and a new profile? UPDATE: I would want to be able to track conversion from a page on the root domain to a sign-up page on the app sub-domain.

    Read the article

  • After a domain change, what can I do to recover lost traffic, rankings, impressions etc? [duplicate]

    - by Felix
    This question already has an answer here: How do I rename a domain and preserve PageRank? 3 answers I moved my site to a legacy exact-match domain I purchased about a couple of months ago. I have seen significant reduction in traffic, impressions, and rankings. I did all the right steps/best practices: change of address in GWT, map old site hierarchy and match to new site for 301 redirects etc. Indexation has gone through the Google process: old site has all but dissappeared from he index and new site is indexed, albeit with some 404 errors which I am addressing. Does anyone else who has gone through eh domain change process have any thoughts/advice? Thanks!

    Read the article

  • Does Google pick up anchor text that is in nested elements?

    - by dangerDAN
    When Google looks at anchor text on a website, will it pickup the text if it is inside nested elements? So for example: <a href="http://www.google.com/">Visit Google</a> To: <a href="http://www.google.com/"> <div class="circle"> <span>Visit Google</span> </div> </a> The reason I ask is because I want to use css3 elements for certain links on my website, to style them as circles. But the anchor text needs to be picked up for these links, so I want to know wether or not the above is bad practice in this case.

    Read the article

  • See how one goal leads to another (i.e., how often downloading the demo leads to purchasing)

    - by s3cur3
    I have two goals set up for my site in Google Analytics: one to download the demo of my software, and the other to buy the full version. I'm having trouble getting statistics on how often the download leads to a purchase. This seems like something the multi-channel funnel is perfect for (as in this question)---I expect that it takes multiple visits for the demo download to convert into an order. However, I can't seem to narrow down the multi-channel funnel data into only those that have one goal followed by another. If it matters, I'm tracking the ordering action as a page view (i.e., someone views the "Thanks for ordering" page) and the demo download as an action (which fires when someone clicks the "download" button). Thanks! - Tyler

    Read the article

  • Why are new pages not being indexed and old pages stay in the index?

    - by ZakGottlieb
    I currently have a site that was recently restructured, causing much of its content to be reposted, creating new URL's for each page. To avoid duplicates, all of the existing pages were added to the robots file. That said, it has now been over a week - I know Google has recrawled the site - and when I search for term X, it is stil the old page that is ranking, with the new one nowhere to be seen. I'm assuming it's a cached version, but why are so many of the old pages still appearing in the index? Furthermore, all "tags" pages (it's a Q&A site, like this one) were also added to the robots a few months ago, yet I think they are all still appearing in the index. Anyone got any ideas about why this is happening, and how I can get my new pages indexed?

    Read the article

  • A drop in SERP after following webmaster guidelines [on hold]

    - by digiwig
    So here's a puzzle for all you SEO gurus out there. I recently launched my own site. I had target keywords which were ranking very well for about 1 month, within the top five and even appearing in first place. In an attempt to maintain good positioning, I followed guidelines by adding robots.txt, an xml sitemap redirecting non-www to www redirecting index.php to root domain adding htaccess 301 redirect for old pages I added rich snippets created a google+ account, verified my picture to appear, I went through each of the webmaster issues with duplicate titles and meta descriptions and improved header tag document outlines i even created a few more blog posts to keep the content freshing and moving. So now my website appears on page 2 with my target keywords - and all because I followed the guidelines. What is happening? I see competitors with stagnant content superglued to position 1.

    Read the article

  • Google active la réponse et l'update partielle dans son Data protocol, pour faciliter la mise à jour

    Google active la réponse partielle et l'update partielle dans son Data protocol, pour faciliter la mise à jour de ses APIs Dans le but de booster la rapidité de ses APIs, Google Data Protocol (qui permet aux développeurs d'écrire des applications liées avec les données contenues dans les produits Google) s'est vu attribué deux nouvelles fonctionnalités cette semaine, au stade expérimental : la réponse partielle, et l'update partielle. Conjointement, ces deux fonctions "peuvent significativement réduire les ressources consommées par le réseau, la mémoire et le CPU" dont on a besoin pour travailler avec les APIs de Google. Pour expliquer le rôle de la réponse partielle, l'équipe du Google Data Protocol donne ...

    Read the article

  • how can i track visits to ALL of the subpages of my website COMBINED TOGETHER?

    - by realcheesypizza
    Right now im using statcounter and google analytics. They are great. But my counts are currently separated. Ex: website.com = 1000 visits a day, website.com/about = 50 visits a day, website.com/privacy = 10 visits a day, etc.. How can have a combined count of all of my sub-pages? (mainpage + about page + about 100 other sub-pages ) I can of course manually add them all together, but that's time consuming because there are many pages. I tried placing a separate tracking code in a php include that sits in each of the sub-pages, but it doesnt seem to be working. It seems to require a single URL to create it, which it then only counts the visits from the one url, rather than ALL of them. Ex: website.com) Any help would be appreciated. Hopefully im just missing something very obvious. Thank you!

    Read the article

  • How to improve a single-paged site search result [closed]

    - by Trisism
    Possible Duplicate: How to SEO a Single-Page website I created an online CV of mine a couple of weeks ago and it has had quite a few visits. Now I want to improve the chance it will appear in google search results; however, my web CV is a one-paged site and it contains only internal links (those with hash #) so I can't really submit a sitemap. I could have changed the internal links to normal links to be processed on server-side, but there's no point of doing so. I'm very new to web SEO so I would really appreciate if somebody can show me what should I do with a single-paged site with internal links to be effectively indexed by crawlers.

    Read the article

  • How to Automate Checking for Stolen Content?

    - by Hisoka
    So I know about tools like Copyscape and Google Alerts.. great tools, but it's quite tedious for me to copy and paste an URL or phrase for every one of my pages in my sites. Is there any tool out there that monitors your website and emails you or alerts you whenever someone has stolen content from your site? The only service I know is CopySentry and honestly, it's too expensive for me since I got thousands of pages I want to monitor... Anyone else have this problem? or is it just me? Thanks for any help.

    Read the article

  • Analytics - Total events divided by number of unique pages?

    - by GeekyAndUnique
    I am using Google Analytics events to track keywords on my articles - not necessarily the best system I know but there are too many for variables I can't easily change it right now - and I would like to be able to see how popular each keyword is by dividing the number of page views with a keyword by the number of unique pages. Is there a/what is the best way of doing this? EDIT FOR CLARITY I currently have a system set up where every time somebody loads an article an event is fired for each of the tags/keywords used, with the keyword being the label. I can currently view my view count for each of the keywords by looking at the total events for each label, however I would like to be able to see which keywords are the most popular by dividing the number of times the event has been fired by the the number of different pages it has been fired from.

    Read the article

  • Which MIME type to compress? and what If I omit the `type` attribute from the HTML?

    - by rockyraw
    Per my request, my webhost had turned mod_deflate ON. In my Cpanel I now have an "Optimize Website" button. Inside that menu I could either choose: "Compress all content" or "Compress the specified MIME types" with the following default MIME types: "text/html text/plain text/xml" Which option should I choose and why? If I choose option 2, which types should I add (is there a recommended list with the exact way they should be written)? According to Google recommendations, I have omitted the type="text/css" attributes from all CSS references, as well as the type="text/javascript" attributes from all script references. Would this hinder the "gzipping" process?

    Read the article

  • Need to add 30K new pages to a 10K page website - troubles ahead? (SEO)

    - by Jurga
    We have a situation with a website where we plan to add a huge amount of new pages. The domain is over 10 years old, approximately 10 thousand indexed pages, and the planned addition is approx. 30K new pages. Any idea how we should go about it? Must we schedule a gradual data release? Have you heard of any industry standards as to how many new pages per day / week / month should be added in order to appear natural and not get in trouble with Google? I.e. should we plan a bi-weekly addition of 5K?

    Read the article

  • Google quitterait la Chine dès le 10 avril et l'annoncerait lundi, selon une source proche du dossie

    Mise à jour du 19 03 2010 par Katleen Google quitterait la Chine dès le 10 avril et l'annoncerait lundi, selon une source proche du dossier Une nouvelle source anonyme vient d'évoquer le possible départ de Google avec, cette fois, une date précise de cessation d'activités. D'après un collaborateur de la firme de Mountaiw View, "Google quitterait la Chine le 10 avril, mais Google n'a pas confirmé l'information pour le moment". Ce témoignage a été reccueilli par CBN (China Business News), et annonce une prise de parole officielle de Google pour expliquer l'organisation de ce départ dès lundi. A suivre donc. Mise ...

    Read the article

  • Google Docs : fin de l'export de documents aux « anciens » formats d'Office .doc .xls et .ppt, mais leur support sera toujours assuré

    Google Docs for Business : fin de l'export de documents aux « anciens » formats d'Office .doc, .xls, et .ppt, mais leur support sera toujours assuré par la suite Pas de nouveauté cette semaine pour les Google Apps for Business, mais une annonce qui intéressera les utilisateurs d'anciennes version de Microsoft Office (antérieures à 2007). La suite hébergée de Google ne permettra plus d'exporter des documents aux « anciens » (sic) formats que sont .doc, .xls, et .ppt. Un changement qui prendra effet dès le 1er octobre. Google précise cependant bien que Google Apps for Business continuera de supporter ces formats et qu'il sera toujours possible d'uploader ce type de doc...

    Read the article

  • l2tp / ipsec debian Openswan U2.6.38 does not connect

    - by locojay
    i am trying to get ipsec/l2tp running on a debian server with an iphone as a client but always get: Dec 2 21:00:04 vpn pluto[22711]: packet from <clientip>:43598: received Vendor ID payload [RFC 3947] method set to=115 Dec 2 21:00:04 vpn pluto[22711]: packet from <clientip>:43598: received Vendor ID payload [draft-ietf-ipsec-nat-t-ike] meth=114, but already using method 115 Dec 2 21:00:04 vpn pluto[22711]: packet from <clientip>:43598: received Vendor ID payload [draft-ietf-ipsec-nat-t-ike-08] meth=113, but already using method 115 Dec 2 21:00:04 vpn pluto[22711]: packet from <clientip>:43598: received Vendor ID payload [draft-ietf-ipsec-nat-t-ike-07] meth=112, but already using method 115 Dec 2 21:00:04 vpn pluto[22711]: packet from <clientip>:43598: received Vendor ID payload [draft-ietf-ipsec-nat-t-ike-06] meth=111, but already using method 115 Dec 2 21:00:04 vpn pluto[22711]: packet from <clientip>:43598: received Vendor ID payload [draft-ietf-ipsec-nat-t-ike-05] meth=110, but already using method 115 Dec 2 21:00:04 vpn pluto[22711]: packet from <clientip>:43598: received Vendor ID payload [draft-ietf-ipsec-nat-t-ike-04] meth=109, but already using method 115 Dec 2 21:00:04 vpn pluto[22711]: packet from <clientip>:43598: received Vendor ID payload [draft-ietf-ipsec-nat-t-ike-03] meth=108, but already using method 115 Dec 2 21:00:04 vpn pluto[22711]: packet from <clientip>:43598: received Vendor ID payload [draft-ietf-ipsec-nat-t-ike-02] meth=107, but already using method 115 Dec 2 21:00:04 vpn pluto[22711]: packet from <clientip>:43598: received Vendor ID payload [draft-ietf-ipsec-nat-t-ike-02_n] meth=106, but already using method 115 Dec 2 21:00:04 vpn pluto[22711]: packet from <clientip>:43598: ignoring Vendor ID payload [FRAGMENTATION 80000000] Dec 2 21:00:04 vpn pluto[22711]: packet from <clientip>:43598: received Vendor ID payload [Dead Peer Detection] Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[4] <clientip> #20: responding to Main Mode from unknown peer <clientip> Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[4] <clientip> #20: transition from state STATE_MAIN_R0 to state STATE_MAIN_R1 Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[4] <clientip> #20: STATE_MAIN_R1: sent MR1, expecting MI2 Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[4] <clientip> #20: NAT-Traversal: Result using draft-ietf-ipsec-nat-t-ike (MacOS X): both are NATed Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[4] <clientip> #20: transition from state STATE_MAIN_R1 to state STATE_MAIN_R2 Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[4] <clientip> #20: STATE_MAIN_R2: sent MR2, expecting MI3 Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[4] <clientip> #20: ignoring informational payload, type IPSEC_INITIAL_CONTACT msgid=00000000 Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[4] <clientip> #20: Main mode peer ID is ID_IPV4_ADDR: '10.2.210.176' Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[4] <clientip> #20: switched from "L2TP-PSK-noNAT" to "L2TP-PSK-noNAT" Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #20: deleting connection "L2TP-PSK-noNAT" instance with peer <clientip> {isakmp=#0/ipsec=#0} Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #20: transition from state STATE_MAIN_R2 to state STATE_MAIN_R3 Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #20: new NAT mapping for #20, was <clientip>:43598, now <clientip>:49826 Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #20: STATE_MAIN_R3: sent MR3, ISAKMP SA established {auth=OAKLEY_PRESHARED_KEY cipher=aes_256 prf=oakley_sha group=modp1024} Dec 2 21:00:04 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #20: Dead Peer Detection (RFC 3706): enabled Dec 2 21:00:05 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #20: the peer proposed: <public ip>/32:17/1701 -> 10.2.210.176/32:17/0 Dec 2 21:00:05 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #20: NAT-Traversal: received 2 NAT-OA. using first, ignoring others Dec 2 21:00:05 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #21: responding to Quick Mode proposal {msgid:311d3282} Dec 2 21:00:05 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #21: us: 171.138.2.13<171.138.2.13>:17/1701 Dec 2 21:00:05 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #21: them: <clientip>[10.2.210.176]:17/61719 Dec 2 21:00:05 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #21: transition from state STATE_QUICK_R0 to state STATE_QUICK_R1 Dec 2 21:00:05 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #21: STATE_QUICK_R1: sent QR1, inbound IPsec SA installed, expecting QI2 Dec 2 21:00:05 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #21: Dead Peer Detection (RFC 3706): enabled Dec 2 21:00:05 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #21: transition from state STATE_QUICK_R1 to state STATE_QUICK_R2 Dec 2 21:00:05 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #21: STATE_QUICK_R2: IPsec SA established transport mode {ESP=>0x05e23c9a <0x216077a9 xfrm=AES_256-HMAC_SHA1 NATOA=10.2.210.176 NATD=<clientip>:49826 DPD=enabled} Dec 2 21:00:26 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #20: received Delete SA(0x05e23c9a) payload: deleting IPSEC State #21 Dec 2 21:00:26 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #20: received and ignored informational message Dec 2 21:00:27 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip> #20: received Delete SA payload: deleting ISAKMP State #20 Dec 2 21:00:27 vpn pluto[22711]: "L2TP-PSK-noNAT"[5] <clientip>: deleting connection "L2TP-PSK-noNAT" instance with peer <clientip> {isakmp=#0/ipsec=#0} Dec 2 21:00:27 vpn pluto[22711]: packet from <clientip>:49826: received and ignored informational message Dec 2 21:00:27 vpn pluto[22711]: ERROR: asynchronous network error report on eth0 (sport=4500) for message to <clientip> port 49826, complainant <clientip>: Connection refused [errno 111, origin ICMP type 3 code 3 (not authenticated)] my setup looks like this verizon fios actiontec -- DMZ-- ddwrt router -- debian xen instance actiontec : 192.168.1.1 ddwrt: 171.138.2.1 debian xen server: 171.138.2.13 forwarded udp 500, 4500, 1701 on ddwrt to debian xen instance. vpn passthrough is enabled /etc/ipsec.conf config setup dumpdir=/var/run/pluto/ nat_traversal=yes virtual_private=%v4:10.0.0.0/8,%v4:192.168.0.0/16,%v4:172.16.0.0/12,%v4:25.0.0.0/8,%v6:fd00::/8,%v6:fe80::/10,%v4:!171.138.2.0/24,%v4:!192.168.1.0/24 protostack=netkey # Add connections here conn L2TP-PSK-NAT rightsubnet=vhost:%priv also=L2TP-PSK-noNAT conn L2TP-PSK-noNAT authby=secret pfs=no auto=add keyingtries=3 # we cannot rekey for %any, let client rekey rekey=no # Apple iOS doesn't send delete notify so we need dead peer detection # to detect vanishing clients dpddelay=30 dpdtimeout=120 dpdaction=clear # Set ikelifetime and keylife to same defaults windows has ikelifetime=8h keylife=1h # l2tp-over-ipsec is transport mode type=transport # left=171.138.2.13 # # For updated Windows 2000/XP clients, # to support old clients as well, use leftprotoport=17/%any leftprotoport=17/1701 # # The remote user. # right=%any # Using the magic port of "%any" means "any one single port". This is # a work around required for Apple OSX clients that use a randomly # high port. rightprotoport=17/%any #force all to be nat'ed. because of ios conn passthrough-for-non-l2tp type=passthrough left=171.138.2.13 leftnexthop=171.138.2.1 right=0.0.0.0 rightsubnet=0.0.0.0/0 auto=route /etc/xl2tp/xl2tp.conf [global] ipsec saref = no listen-addr = 171.138.2.13 ;port = 1701 ;debug network = yes ;debug tunnel = yes ;debug network = yes ;debug packet = yes [lns default] ip range = 171.138.2.231-171.138.2.239 local ip = 171.138.2.13 assign ip = yes require chap = no refuse pap = no require authentication = no ;name = OpenswanVPN ppp debug = yes pppoptfile = /etc/ppp/options.xlt2tpd lenght bit = yes /etc/ppp/options.xl2tpd ;require-mschap-v2 pcp-accept-local ipcp-accept-local ipcp-accept-remote ;ms-dns 171.138.2.1 ms-dns 192.168.1.1 ms-dns 8.8.8.8 name l2tpd noccp auth crtscts idle 1800 mtu 1410 mru 1410 lock proxyarp connect-delay 5000 debug dump logfd 2 logfile /var/log/xl2tpd.log ipsec verify Checking your system to see if IPsec got installed and started correctly: Version check and ipsec on-path [OK] Linux Openswan U2.6.38/K3.0.0-1-amd64 (netkey) Checking for IPsec support in kernel [OK] SAref kernel support [N/A] NETKEY: Testing XFRM related proc values [OK] [OK] [OK] Checking that pluto is running [OK] Pluto listening for IKE on udp 500 [OK] Pluto listening for NAT-T on udp 4500 [OK] Two or more interfaces found, checking IP forwarding [FAILED] Checking NAT and MASQUERADEing [OK] Checking for 'ip' command [OK] Checking /bin/sh is not /bin/dash [WARNING] Checking for 'iptables' command [OK] Opportunistic Encryption Support [DISABLED] The failed can be ignored i guess since cat /proc/sys/net/ipv4/ip_forward returns 1 any help would be much appreciated as i don't have any idea why this is not working

    Read the article

< Previous Page | 287 288 289 290 291 292 293 294 295 296 297 298  | Next Page >