Search Results

Search found 19155 results on 767 pages for 'url redirection'.

Page 141/767 | < Previous Page | 137 138 139 140 141 142 143 144 145 146 147 148  | Next Page >

  • What's the easiest way to get the result of an HTTP GET request in using URL in JRuby

    - by sipwiz
    I'm attempting to build a Tropo Ruby application and I need to retrieve the result of an HTTPS GET. The Tropo platform doesn't have the httpclient Ruby gem so I can't use that. The Ruby engine used is JRuby so a suggestion has been to make use of the Java URL class to do the request. I've played around with it a little bit and I seem to be able to create the URL object ok but am now struggling with how to get the results of executing the request. How do I do it? javaURL = java.net.URL.new svcURL transferResult = javaURL.getContent()

    Read the article

  • Safe non-tamperable URL component in Perl using symmetric encryption?

    - by Randal Schwartz
    OK, I'm probably just having a bad Monday, but I have the following need and I'm seeing lots of partial solutions but I'm sure I'm not the first person to need this, so I'm wondering if I'm missing the obvious. $client has 50 to 500 bytes worth of binary data that must be inserted into the middle of a URL and roundtrip to their customer's browser. Since it's part of the URL, we're up against the 1K "theoretical" limit of a GET URL. Also, $client doesn't want their customer decoding the data, or tampering with it without detection. $client would also prefer not to store anything server-side, so this must be completely standalone. Must be Perl code, and fast, in both encoding and decoding. I think the last step can be base64. But what are the steps for encryption and hashing that make the most sense?

    Read the article

  • Does the android market use your merchant API callback URL during the Google checkout process for pa

    - by tootflow
    If I have a paid app on Android Market, will my merchant API callback URL get a hit when the customer goes to check out? I am wondering how merchants/vendors/developers secure their applications for android market. I have used google checkout before to implement a custom integration, so I understand how that all works. What I do not see anywhere that I have looked is what, if any, integration does the merchant/developer have with the Android Market checkout process? I understand that the market uses google checkout, but in that case the market hosts the checkout process and not the merchant, so it's not clear whether they give your URL a call. So where is the hook? Am I right in assuming it is the merchant API callback URL?

    Read the article

  • Possible to link an image in a playlist to a URL, with Flowplayer?

    - by Yegor
    I have a simple Flowplayer playlist. First one plays a short clip, which then cuts off, and an image is displayed. Is there a way to make the image clickable? here is what I have: <a style="display:block;width:640px;height:360px" id="player"> </a> <script> flowplayer("player", "../flowplayer-3.1.5.swf", { playlist: [ {url: 'http://e1h13.simplecdn.net/flowplayer/flowplayer.flv', duration: 10}, {url: 'http://www.domain.com/image.jpg', duration: 100000}, ], plugins: { controls: null } }); </script> If there is another way of doing this, Im all ears. I need a video to play for 10 seconds (or entirely, since the file is gonna be 10 seconds long), and then display some text + a URL.

    Read the article

  • How do I verify the URL that ActiveResource intends to use on a call?

    - by btelles
    Hi there, I'm trying to request data from another Rails app, but I keep getting a "URL not found" error. Is it possible to use the debugger to figure out the intended URL and parameters for an intended ActiveResource call? For example, I'd like to make a GET request with the following parameters: https://some_server.com/employees/search.xml?last_name=smith And I have the following ActiveResource model class Employee < ActiveResource::Base self.site = "https://.some_server.com" end And the some_server.com app has a search action that maps this way: map.connect "/employees/search.:format", :controller => 'employees', :action => 'search' But when I try this: Employee.new(:last_name => 'smith).get(:search) I get a 404 "Service not found" error. To me, that sounds like the URL that ACtiveResource is requesting is incorrect, 'cause when I make the above request via a browser, I get the xml. Any idea what I'm doing wrong? Cheerio, Berns

    Read the article

  • What parameters should I use in a Google Maps URL to go to a lat-lon?

    - by xpda
    I would like to produce a url for Google Maps that goes to a specific latitude and longitude. Now, I generate a url such as this: http://maps.google.com/maps?z=11&t=k&q=58 41.881N 152 31.324W The resulting map comes up with a round "A" balloon pointer, which seems to point to the nearest named object, and a green arrow, which points to the lat-lon. Sometimes, as in this example, the "A" pointer is centered and is far enough away that you cannot see the pointer to the lat-lon. (Zoom out to see both pointers in this example. The "A" pointer is in the center of Alaska, while the lat-long pointer is on Kodiak Island.) Are there some parameters I can use in the Google Maps URL that will produce a single pointer to a designated lat-lon? (This loads in a separate window. It is not embedded.)

    Read the article

  • Rewrite URL - how to get the hostname and the path?

    - by buggy1985
    Hello, I want to use mod_rewrite to hide the URL, that generates my XML file. So if this is the requested URL: http://www.domain.com/path/to/ page ? cid=8001&srt=po This URL should be executed: http:// www.xmldomain.com/bla/page ? cid=8001&srt=po &rtype=xslt&xsl=http://www.domain.com/path/to/ page .xsl* $1 should be http://www.domain.com/path/to/ (How can I do this?) $2 should be page $3 should be cid=8001&srt=po Thanks!

    Read the article

  • Do Url shorteners such as bit.ly serve any purpose in web development?

    - by gath
    I might be missing something here but am unable to understand the true purpose of link shortener services like bit.ly or is it just a lot of hype! Truly if you type nyti.ms on your URL bar on the browser it takes you to www.nytimes.com, but still what is so hard in typing "nytimes.com! which is far much comprehensible to read and understand leave alone to remember! Does shortening URL serve any purpose in web development? Why should someone want use obscure shortened URL?

    Read the article

  • Is it possible to maintain the url while redirecting to a classic asp page from a controller?

    - by Matthias
    While migrating a site from a classic asp to MVC, I'm having the problem that not all controllers are implemented yet. For those which are not implemented, I'd like to serve the classic asp page (say /product.asp?id=123) while maintaining the nice url /product/123. To accomplish this I implemented a dummy ProductController which returns a RedirectResult to the classic asp url. But that changes the url in the browsers navigation bar. Requirement has it, that the urls should always be a clean (mvc) one, eventhough the page has not yet been fully migrated. If this can't be done using a dummy controller, what would be an alternative option to solve this problem? Thanks in advance!

    Read the article

  • change the return of to_param by action? or, terribly broken edit action if perma-url is returned in

    - by Stacia
    I did some stuff to have to_param return a perma-url so I'd have seo friendly links. Upon creation and update it will generate and return the perma-url in a to_param method that I put in the model. However this causing me some grief in other areas. I have a form that looks like this: <% @apps.each do |app| %> <% fields_for "[id][]", app do |fields| %> .... this fails terribly since it tries to insert the perma-url in place of the id. If I could just restrict to_param by action, that could solve things. If I change app to app.id, that fails since I have stuff like this: fields.check_box :is_featured . I suppose I could get the app instance, but that feels hacky. Again please excuse my lack of rails-foo ;)

    Read the article

  • How to validate a bunch of proxies against a URL?

    - by NJTechGuy
    I have a list of 100 proxies. The URL I am interested in is abc.com. I want to check the number of proxies which can successfully fetch this URL and the time taken for the same. I am hoping I made sense. I am a Python noob. I am looking for a code snippet. A helping hand is really appreciated :) Proxies : 200.43.54.212 200.43.54.212 200.43.54.212 200.43.54.212 URL : abc.com Desired result : Proxy isGood Time 200.43.54.112 n 23.12 200.43.54.222 n 12.34 200.43.54.102 y 11.09 200.43.54.111 y 8.85 p.s : All the above proxies have ports either 80 or 8080

    Read the article

  • Using a custom URL parameter in Wordpress (with permalinks)?

    - by kiko
    Hello - Is there a way to perhaps edit .htaccess to add a custom URL parameter to Wordpress, so that the parameter is not stripped out by permalinks? I have a Wordpress site with a page that queries a separate (non-Wordpress) database. I passed the URL parameter "pubID" to display individual books and it is working OK. Example: http://www.uglyducklingpresse.org/catalog/browse/item/?pubID=63 But the individual books are not showing up properly in Google - maybe because they all have the same auto-generated "canonical" URL meta tab - one with the "pubID" parameter stripped out. Thank you for any help.

    Read the article

  • How can I view the all inherited url-resolution rules affecting a given directory?

    - by john.designop.us
    I work on two sites hosted on the same server, using the same CMS configurations and identical .htaccess files in their respective document roots. One site is letting me use the CMS's clean-url mode, and the other isn't. Site #2 functions fine in ?=messy-url mode, but when I turn clean urls on in the admin panel, and request a rewritten URL, I get a 404 error served before the CMS sees the request. I've contacted the server administrator, but he isn't inclined to provide support and the site owners are beholden to this hosting provider. I have shell access to the Linux-based server, and I can verify that mod_php and mod_rewrite are active, but I don't know what more I can do to troubleshoot this issue. Is there any way to identify directives upstream that may be differentiating the way http requests are handled by the two sites? Thanks!

    Read the article

  • How do I create a no-javascript url hash handler for my website?

    - by Kenny Bones
    Ok, I'm not sure how this is normally done. But I've got a script that basically empties a div of content and then loads content from a div from a separate webpage, without reloading the current page. This works great. It's taken from this example actually, from net tuts (great site btw) http://nettuts.s3.amazonaws.com/011_jQuerySite/sample/index.html And the guy who wrote this even though about handling the url's since the url don't change when using his method. So he wrote a javascript snippet that looks up the url and loads the content accoringly. Which is not working btw. But I was thinking about people who don't have javascript enabled, or iPhone and iPad users ;) Copying URLs and sending to a friend won't work at all. So how is this typically done? And can it be done without javascript? Possibly by php?

    Read the article

  • Getting URL of executing JavaScript file (IE6-7 problem mostly)...

    - by TooTallNate
    Hey all, I've been trying to throw together a generic function that retrieves the absolute URL of an executing JavaScript file on a web page: http://gist.github.com/433486 It works great in almost all the browsers I've tested (Firefox, Chrome, Safari, Opera v10 at least, and IE 8). It seems to fail, however, in IE 6 and 7. The callback function gets executed, but the retrieved name is simple the URL to the main HTML page, not the JavaScript file. So all I'm really asking is if there's some other way of getting the URL of the current JavaScript file (which could be IE 6 and 7 specific)? Thanks in advance!

    Read the article

  • files end up in wrong directory after svn copy

    - by Andy
    I want to copy the trunk of one project to another, so I use the following command: svn copy -m "test" (url of project from)/Trunk/ (url of project to)/Trunk/ The files ended up in (url of project to)/Trunk/Trunk/ instead of (url of project to)/Trunk/ where I want them to go. Have I done something wrong? What should I do if I do if I want the files in (url of project to)/Trunk/? Doing the the following does not seem to make sense: svn copy -m "test" (url of project from)/Trunk/ (url of project to)/

    Read the article

  • remove default.aspx from a request

    - by b0x0rz
    i am trying to remove default.aspx from any request that might have it. protected void Application_BeginRequest(object sender, EventArgs e) { HttpContext context = HttpContext.Current; string url = context.Request.Url.ToString(); // remove default.aspx if (url.EndsWith("/default.aspx", StringComparison.OrdinalIgnoreCase)) { url = url.Substring(0, url.Length - 12); } context.Response.Redirect(url); } gives an error: **too many redirects occurred trying to open...** what can i change to make it work? thnx

    Read the article

  • Regex not being greedy enough

    - by Chad
    I've got the following regex that was working perfectly until a new situation arose ^.*[?&]U(?:RL)?=(?<URL>.*)$ Basically, it's used against URLs, to grab EVERYTHING after the U=, or URL= and return it in the URL match So, for the following http://localhost?a=b&u=http://otherhost?foo=bar URL = http://otherhost?foo=bar Unfortunately an odd case came up http://localhost?a=b&u=http://otherhost?foo=bar&url=http://someotherhost Ideally, I want URL to be "http://otherhost?foo=bar&url=http://someotherhost", instead, it is just "http://someotherhost"

    Read the article

  • MVC4 App opens with directory listing and gives a 404 for any direct URL's entered in the browser

    - by ProfK
    I've just deployed a previously (on my local IIS) working MVC4 app to IIS 7.5 on the dev server. After tweaking this and that - one knows how these things get forgotten - the app finally launches, but shows a directory listing of the app root. Clicking on most links there works, opening the directory listing of the sub-directory. Elmah logs no errors and /elmah.asd also gives a 404. The site has an appropriate localhost binding in the hosts file. I can find nothing wrong. MVC is installed on the server, as another MCV app works fine.

    Read the article

  • Tracking download of non-html (like pdf) downloads with jQuery and Google Analytics

    - by developerit
    Hi folks, it’s been quite calm at Developer IT’s this summer since we were all involved in other projects, but we are slowly comming back. In this post, we will present a simple way of tracking files download with Google Analytics with the help of jQuery. We work for a client that offers a lot of pdf files to download on their web site and wanted to know which one are the most popular. They use Google Analytics for a long time now and we did not want to have a second interface in order to present those stats to our client. So usign IIS logs was not a idea to consider. Since Google already offers us a splendid web interface and a powerful API, we deceided to hook up simple javascript code into the jQuery click event to notify Analytics that a pdf has been requested. (function ($) { function trackLink(e) { var url = $(this).attr('href'); //alert(url); // for debug purpose // old page tracker code pageTracker._trackPageview(url); // you can use the new one too _gaq.push(["_trackPageview",url]); //always return true, in order for the browser to continue its job return true; } // When DOM ready $(function () { // hook up the click event $('.pdf-links a').click(trackLink); }); })(jQuery); You can be more presice or even be sure not to miss one click by changing the selector which hooks up the click event. I have been usign this code to track AJAX requests and it works flawlessly.

    Read the article

  • Insufficient Permissions Problems with MSDeploy and TFS Build 2010

    - by jdanforth
    I ran into these problems on a TFS 2010 RC setup where I wanted to deploy a web site as part of the nightly build: C:\Program Files (x86)\MSBuild\Microsoft\VisualStudio\v10.0\Web\Microsoft.Web.Publishing.targets (3481): Web deployment task failed.(An error occurred when reading the IIS Configuration File 'MACHINE/REDIRECTION'. The identity performing the operation was 'NT AUTHORITY\NETWORK SERVICE'.)  An error occurred when reading the IIS Configuration File 'MACHINE/REDIRECTION'. The identity performing the operation was 'NT AUTHORITY\NETWORK SERVICE'. Filename: \\?\C:\Windows\system32\inetsrv\config\redirection.config Error: Cannot read configuration file due to insufficient permissions  As you can see I’m running the build service as NETWORK SERVICE which is quite usual. The first thing I did then was to give NETWORK SERVICE read access to the whole directory where redirection.config is sitting; C:\Windows\system32\inetsrv\config. That gave me a new error: C:\Program Files (x86)\MSBuild\Microsoft\VisualStudio\v10.0\Web\Microsoft.Web.Publishing.targets (3481): Web deployment task failed. (Attempted to perform an unauthorized operation.) The reason for this problem was that NETWORK SERVICE didn’t have write permission to the place where I’ve told MSDeploy to put the web site physically on the disk. Once I’d given the NETWORK SERVICE the right permissions, MSDeploy completed as expected! NOTE! I’ve not had this problem with TFS 2010 RTM, so it might be just a RC issue!

    Read the article

  • Can't access IIS 7 server URL from the same IIS 7 server.

    - by Kevin Raffay
    We have an intranet site ie, xxx.yyyy.com, that users access by entering "http"://xxx.yyy.com. Our problems started when we migrated to IIS 7 running on a new 2003 server. We got rid of our single-sign on code and implemented a security model where we capture a user's domain credentials which we then authenticate against a DB. In order to get the domain credentials passed to our ASP.NET app, we have the following settings: Anonymous Authentication:Disabled ASP.NET Impersonation: Enabled Basic/Digest/Forms Authentication: Disabled Windows Authentication: Enabled We allow "*" and deny "?" in the web.config. Browsing "http"://xxx.yyy.com from any client PC results in a domain login prompt, and if your enter a proper user/pwd, you can get in. However, browsing "http"://xxx.yyy.com while remoting into the server results in 3 domain login prompts and eventually a 401 error - unauthorized. We have traced this behavior to problems with our web site where we have pages doing "screen scraping" using the HttpRequest calling a url on the same server. When doing a HttpRequest from any other client, using a test harness that passes authorized credentials, all is good. So internal HttpRequest calls on the server fail, just like attempts to browse that server's url from within a remote session. Why would a to "http"://xxx.yyy.com on server xxx.yyy.com fail authentication?

    Read the article

  • How to remove that extra text from URL titles in Firefox browser?

    - by amar
    I am using Firefox with Vimperator. On Mac I use Shiori to bookmark. I also use Pinboard's default add-on (on both Mac and Windows, but mainly on Windows as there's no Shiori). When I used to bookmark, I had this behaviour[A] (before changing Vimperator options): Example: URL: xyz.com TITLE: xyz website Pinboard add-on would fetch the title as xyz website which was fine (I reckon it directly fetches TITLE form the URL). But after installing and starting to use Shiori the title I was getting in its "title" field was xyz website - Vimperator exactly the same what I could see in tabs (seems it's getting what the browser is feeding it). So, I :set titlestring= in Vimperator to remove that extra Vimperator from title. Now, this is what I am getting [B]: When I try to bookmark using Pinboard add-on the title is undefined Using Shiori it results in xyz website - undefined. I tried to change it back to original[C] :set titlestring=Mozilla Firefox (assuming this was the original, before Vimperator) but the results are still the same as [B]. How to get read of that extra - undefined or that extra - Vimperator or - Mozilla Firefox for that matter, while bookmarking pages either with Pinboard add-on or Shiori? One workaround is, instead of no string just add a white-space while setting title in Firefox via Vimperator there but that results in - appended to titles.

    Read the article

  • .htaccess non-www to www rule seems to work but the URL isn't changing in the address bar

    - by SnakeByte
    On a joomla site, apache, shared hosting, I'm using next .htaccess rule: RewriteCond %{HTTP_HOST} !^www\. RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1 [R=301,L] The problem is that the browser's address bar text does not change from example.com to www.example.com. It seems the redirect actually works because all the links on the pages are changed to www. And after clicking on any link from there it always continues to have www added. The problem is the first page (no matter which one) that is loaded using browser's address bar - like example.com or example.com/random-page. Solved. The redirect actually works.

    Read the article

  • Do you need to crawl the whole internet to find backlinks of a URL?

    - by Luca Matteis
    Say I want to retrieve all the sites on the web that have a specific link on them. For example I want to know all the backlinks made to my blog, on other websites. There are services out there that do this: http://www.backlinkwatch.com/index.php - was wondering how they achieve this functionality. Is crawling the entire internet the only option or are there third-party ways of doing this, say using Google.

    Read the article

< Previous Page | 137 138 139 140 141 142 143 144 145 146 147 148  | Next Page >