Search Results

Search found 8013 results on 321 pages for 'clean urls'.

Page 174/321 | < Previous Page | 170 171 172 173 174 175 176 177 178 179 180 181  | Next Page >

  • What's the easiest way to delete Vim swapfiles I've already recovered from?

    - by Peeja
    Sometimes Vim crashes and leaves me a swapfile. That's awesome. Then I open the file I was editing and Vim asks me if I want to recover. I do, thanks. When it's done, Vim tells me, You may want to delete the .swp file now. Why, yes, I do. How do I do that? I figured it would just start using the old swapfile as a swapfile again and clean it up when I quit, but that's not true. It makes a new one, cleans that one up, and when I open the file again it prompts me again to recover from the first one. Surely I'm missing something.

    Read the article

  • opening iWorks documents in iPad UIWebView

    - by user369156
    Hello, I'm writing an iPad application that has a UIWebView which I open word and excel documents in, but I want the user to be able to import those documents into the iWorks applications, Pages and Numbers, just like how you can do it in Safari if you open a document. If you open a document in Safari on the iPad, there'll be a button on the top bar that says "Open in..." and you can choose applications to open in. You get the top bar to appear by tapping on middle of the page. So is there an option you can set to allow UIWebView to show up the bar and automatically detect the content type and populate the list with applications you can import in? Or do I have to build this myself? And if I have to build my own, how do I open URLs to import documents into Pages and Numbers etc? Thanks, -David

    Read the article

  • Windows 2008 R2 Not resolving IP's or domains

    - by InsaneAR
    I have two servers with Microsoft Windows Server 2008 R2 x64, and today I realized that one of them don't resolve IP's or domain names. If I ping the same machine IP, it respond, but if I ping an external IP (like 8.8.8.8 or any other) it does not work. The same with domain names. At Windows Event Viewer I'm having Warning 1014: Name resolution for the name timed out after none of the configured DNS servers responded THE SERVER HAS CONNECTIVITY, I'm connected remotely and the services and incoming traffic (like FTP, etc) are working OK. Things I've done & tested: Run Kaspersky Virus Removal 10 Restart/Turn off the DNS Client service Clean the hosts file Reboot the server Turn off the Windows Firewall (there's not another firewall) Quit or replace the current DNS servers I just can ping the same machine IP or the gateway! Can anybody help me?

    Read the article

  • guide on crawling the entire web ?

    - by bohohasdhfasdf
    i just had this thought, and was wondering if it's possible to crawl the entire web (just like the big boys!) on a single dedicated server (like Core2Duo, 8gig ram, 750gb disk 100mbps) . I've come across a paper where this was done....but i cannot recall this paper's title. it was like about crawling the entire web on a single dedicated server using some statistical model. Anyways, imagine starting with just around 10,000 seed URLs, and doing exhaustive crawl.... is it possible ? I am in need of crawling the web but limited to a dedicated server. how can i do this, is there an open source solution out there already ? for example see this real time search engine. http://crawlrapidshare.com the results are exteremely good and freshly updated....how are they doing this ?

    Read the article

  • Retrieving a page of that has a redirect

    - by Dmitry Makovetskiyd
    I get my page content with this function: private function fetch_url($url){ $ch=curl_init($url); curl_setopt($ch, CURLOPT_RETURNTRANSFER, true); curl_setopt($ch, CURLOPT_TIMEOUT, 320); $this->doc = curl_exec($ch); $this->status_code= curl_getinfo($ch, CURLINFO_HTTP_CODE); // echo $this->doc; curl_close($ch); } The problem is that with some urls dont exist on a webpage and there is a redirect to another page.. So say if I put the parameter: http://example.com/uncategorized/ It redirect me to : http://example.com/mature/ The problem is with curl, I dont get any content.. But my aim is to get the content of that page redirect.. Is there an easy way to get the function to work in the way I want..?

    Read the article

  • Unknown device error installing Lexmark printer in Vista

    - by packotjimmy
    I have been using Lexmark X3650 All-In-One on my Windows Vista Desktop PC for a year without problem. Yesterday, I was prompted to plug in the USB from the printer and after that I got errors: "The USB device has malfunctioned." ,"The unknown device has been installed unsuccessfully.". Then, the printer just stoped responding. I have uninstalled the driver from Device Manager and software (using Lexmark's clean utility), rebooted, reinstalled, same thing, tried different ports/cable. Vista displays these two errors stated above immediately upon plug in. Please help me!

    Read the article

  • Automating SSRS Report Deployment

    - by MikeTWebb
    Hello....I have a web site in 3 environments (DEV/TEST/PROD). The Web Site uses SSRS which also has 3 different environments(DEV/TEST/PROD). Currently, I am deploying the reports using the Manual Deployment mechanism provided in the SSRS Solution (i.e., right click a report and click "Deploy"). I have to manually change the Deployment URLs for each environment. Is there a way to Automate this so that reports are automatically deployeed to PROD and TEST if a change occurs in DEV? I've seen some info on RSScripter but it looks a little tricky/clunky...I was hoping for a more succinct/streamlined solution. Thanks MikeTWebb

    Read the article

  • Regular Expression - Capture and Replace Select Sequences

    - by Chad
    Take the following file... ABCD,1234,http://example.com/mpe.exthttp://example/xyz.ext EFGH,5678,http://example.com/wer.exthttp://example/ljn.ext Note that "ext" is a constant file extension throughout the file. I am looking for an expression to turn that file into something like this... ABCD,1234,http://example.com/mpe.ext ABCD,1234,http://example/xyz.ext EFGH,5678,http://example.com/wer.ext EFGH,5678,http://example/ljn.ext In a nutshell I need to capture everything up to the urls. Then I need to capture each URL and put them on their own line with the leading capture. I am working with sed to do this and I cannot figure out how to make it work correctly. Any ideas?

    Read the article

  • Getting rid of index.php in the URL when using recess framework and lighttpd

    - by spudnik1979
    I am using the recess php framework with lighttpd Does anyone know how I can use the shorter urls of: http://www.myserver.com/recess Instead of: http://www.myserver.com/index.php/recess The recess readme file says that if I have mod_rewrite I can use the shorter url: -- "Do you have mod_rewrite? -- Yes: Open your browser to the location you unzipped -- No: Open your browser to the location you unzipped followed by index.php" I do have mod_rewrite enabled on lighttpd and i have removed the index.php but I get a 404. Do I need any special rules in my lighttpd.conf?

    Read the article

  • Disk Cleanup on Server 2008 R2 is ineffective

    - by cparker4486
    I have a user profile with ~2.9GB of Error Reports backed up in the ReportQueue folder (C:\Users\UserName\AppData\Local\Microsoft\Windows\WER\ReportQueue). Running DiskCleanup as the administrator does not detect these files and therefore does not clean them up. However, running the utility as the user shows an even larger amount (12.4GB!) of error reporting files. As seen below: The problem is that after running the cleanup utility the disk spaced used does not decrease by anywhere near 12.4GB and running the utility again detects the same 12.4GB of files. What is the problem here? Alternatively, can I manually delete the files in the ReportQueue folder?

    Read the article

  • Belkin Wireless G router not working on Windows 7

    - by coore
    I have a Belkin F5D7230-4 Wireless G router which used to work on XP. I have a clean Windows 7 install (on the same desktop PC), and it just won't work. Device Manager lists the adapter as a USB device, but it says "The drivers for this device are not installed", even though the driver installer said the installation was OK. Manage Wiresless Networks says "wireless is not currently enabled". The WLAN AutoConfig service is running. My specific question is which of the two symptoms above should I go for first? Should I try to fix the driver installation somehow, or should I try to enable wireless? Is one caused by the other, or is this two independent problems? Any other thoughs that could help me are also appreciated.

    Read the article

  • Why does Perl lose foreign characters on Windows input - can this be fixed (if so, how) or is Perl an outdated dinosaur that just can't handle this?

    - by Alex R
    Note below how ã changes to a This is causing me a huge problem as foreign characters show up in URLs, e.g. http://pt.wikipedia.org/wiki/Cão The OS is Windows 7, 64-bit. The Perl is: This is perl 5, version 12, subversion 2 (v5.12.2) built for MSWin32-x64-multi-thread (with 8 registered patches, see perl -V for more detail) Copyright 1987-2010, Larry Wall Binary build 1202 [293621] provided by ActiveState http://www.ActiveState.com Built Sep 6 2010 22:53:42 Additional update: To get around my particular problem, I tried using File::Find instead of piped input. The issue actually gets worse:

    Read the article

  • Remove special chars from URL

    - by John Jones
    Hi, I have a product database and I am displaying trying to display them as clean URLs, below is example product names: PAUL MITCHELL FOAMING POMADE (150ml) American Crew Classic Gents Pomade 85g Tigi Catwalk Texturizing Pomade 50ml What I need to do is display like below in the URL structrue: www.example.com/products/paul-mitchell-foaming-gel(150ml) The problem I have is I want to do the following: Remove anything with braquets(and the braquets) Remove any numbers next to g or ml e.g. 400ml, 10g etc... I have been banging my head trying different string replaces but cant get it right, I would really appreciate some help. Cheers

    Read the article

  • When should I upgrade to Ubuntu 10.04 (Lucid Lynx)? [closed]

    - by Emyr
    I'm a web developer for a small non-IT firm. When 9.10 came out, I was using it with no adverse effects from about a month before release (iirc, first beta), initially as an upgrade but as a clean install later to ensure my system would be consistent with most other 9.10 systems. The last alpha of 10.04 came out last week, with another 2 weeks before beta. I'm quite eager to do it today, but obviously the usual "not for production systems" notice is still in place. When should I upgrade? Do I need to worry about software installed from source? (./configure, make, make install etc) Is the attraction of a non-brown theme really this tempting for you?

    Read the article

  • Black spray (?) on a laptop's heat sink assembly

    - by arnehehe
    We have a HP Pavilion DV3 4000SB that was given an error on startup that the fan was having problems (System Fan 90b error message) So I decided to open the laptop up and clean the fan, as that seemed the easiest option. Upon opening it up, at first I froze and thought "No wonder! The whole thing's burnt!", but then I started thinking it looked more like someone used a spraycan on it. I looked up a video on youtube of someone disassembling his Pavilion and noticed there wasn't any of the black on his, so my question here is: Have any of you encountered this, or know what it is? Or if it's something that could point to a problem?

    Read the article

  • Why are awkward numbers and letters in my query string??

    - by Nicole
    Hi everyone!! I have a MVC website and when I execute it, the query string (url) that appears is: http://localhost:6970/(S(51aegi45qneolxa0oxwuzh55))/default.aspx what the hell is that strange combination of numbers and letters??? please help!!!!! previously i had been working with the session expires to redirect to one of my views, could be something of that?? i have added the following action filter: public override void OnActionExecuting(ActionExecutingContext filterContext) { HttpContextBase ctx = filterContext.HttpContext; // check if session is supported if (ctx.Session != null) { // check if a new session id was generated if (ctx.Session.IsNewSession) { // If it says it is a new session, but an existing cookie exists, then it must // have timed out string sessionCookie = ctx.Request.Headers["Cookie"]; if ((null != sessionCookie) && (sessionCookie.IndexOf("ASP.NET_SessionId") >= 0)) { filterContext.HttpContext.Response.Redirect("~/Error/SesionExpirada", true); } } } base.OnActionExecuting(filterContext); } but if I put it in ignore it stills add that rare thing to my urls. Please help!!! thanks

    Read the article

  • Add RESTful Action

    - by Drew Rush
    The source of my information is section 2.9 here: [http://guides.rubyonrails.org/routing.html#connecting-urls-to-code][1] What I'm trying to do is add a custom action "search" and corresponding view. So, as it says to do in the documentation, I've added this code in my config/routes.rb file: resources :dimensions do collection do get "search" end end I've also defined in the dimensions_controller file: def search @dimensions = Dimension.all respond_to do |format| format.html # search.html.erb format.json { render json: @dimensions } end end I then stopped and restarted the rails server, but when I navigate to /dimensions/home, I'm still getting this error message: Couldn't find Dimension with id=search Also showing that my parameter is: {"id"=>"search"} So am I just missing another bit of code that gives the instruction to interpret /dimension/search as a collection action as opposed to the show action? Thanks for your time.

    Read the article

  • License for website article

    - by queueoverflow
    On my personal website, I have some technical ideas and some source code snippets that I share with everybody. To make it clean that everyone can use those snippets as they like, as long as I do not have to provide any warranty, I would like to add a license to the some of the texts. The bigger programs come with GPLv2+, which I think is a reasonable license for free code. Does it make sense to use the MIT License or the GNU Free Documentation License for these texts or should I just go with CC-BY? I am a German citizen, so I heard that the American licenses do not really apply to me at all. If so, that would be another advantage for the Creative Commons Family of License.

    Read the article

  • Is there a chroot build script somewhere?

    - by Nils
    I am about to develop a little script to gather information for a chroot-jail. In my case this looks (at the first glance) pretty simple: The application has a clean rpm-install and did install almost all files into a sub-directory of /opt. My idea is: Do a find of all binaries Check their library-dependencies Record the results into a list Do a rsync of that list into the chroot-target-directory before startup of the application Now I wonder - ist there any script around that already does such a job (perl/bash/python)? So far I found only specialized solutions for single applications (like sftp-chroot). Update I see three close-votes for the reason "off topic". This is a question that arose because I have to install that ancient piece of software on a server at work. So if you still feel this is off-topic - leave a comment...

    Read the article

  • How to install Windows on a laptop with no CDROM drive?

    - by Jason Kester
    I have an old Thinkpad X60 that I'd like to wipe clean and rebuild. Seeing as this machine doesn't have an optical drive, what's the easiest way of installing Windows XP? I have an external USB hard drive available. Would it be possible to run the install from that instead? Otherwise, what options do I have? Edit: assuming we're using a USB mass storage device... Is there a BIOS setting that I would need to change, or will it configure itself automatically? Would the USB drive need to be configured in any special manner, or would simply having a copy of the Windows CD files in a directory there be sufficient? Since the first couple answers that came in were basically "yes", I guess I didn't phrase my question correctly. I'm asking for detailed instructions on how to do this, not just a sanity check that I'm headed in the right direction. Thanks!

    Read the article

  • How can I find the names of AD Group policies that a user/pc is using?

    - by Russ
    I am having trouble locating some settings in group policy so I can make changes due to the convoluted nature of our policies. What I would like to be able to do is go to a specific PC and see what group policies are being applied, so I can focus on those policies. My goal would be to clean up the GP's a bit, while allowing me to "walk the tree" to see what people have implemented and what is worthless. Thanks. EDIT: In this specific case, I am looking to find which GP maped drives are configured in. (User Configuration -- Preferences -- Windows Settings -- Drive Maps)

    Read the article

  • Can I delete the OEM partition on the new Dell XPS 15?

    - by timepilot
    My new Dell XPS 15 L521X just arrived. I need to set this up to dual boot Linux. Sadly, the system comes with four primary partitions. I can't do a clean install at the moment, so one of the partitions will have to be deleted before I can install Linux. The layout is as follows: OEM: 39mb Hibernation: 8gb OS: 457gb Recovery: 12gb Obviously, I can't delete the Hibernation and OS partitions (will shrink to make space for Linux) and I'd like to keep the recovery partition if possible. So my question is what is on the small OEM partition? What functionality will I lose if I delete it?

    Read the article

  • Why FILTER_VALIDATE_URL return FALSE for only this url?

    - by Axel
    Hi, i have the following code: <?php $pictureurl="http://icons3.iconfinder.netdna-cdn.com/data/icons/pool/poolbird.png"; if(filter_var($pictureurl, FILTER_VALIDATE_URL) === FALSE){ echo "Invalid Url"; exit; }else{ echo "Works!"; } ?> This display "invalid url" for the above url, but not for other simpler urls. Is this a bug? you can even access the image. And the most important is what's the solution for this? Thanks

    Read the article

  • root path for multiple controllers on rails routes

    - by Lee
    I have two resource controllers where I am using a slug to represent the ID. (friendly_id gem). I am able to have the show path for one resource on the route but not for two at the same time. ie. root :to => 'home#index' match '/:id' => "properties#show" match '/:id' => "contents#show" Basically I want urls like, # Content domain.com/about-us domain.com/terms # Property domain.com/unique-property-name domain.com/another-unique-property-name Whatever resource I put on top works. Is there a way to do this? Thanks in advace if you can help.

    Read the article

  • What's the difference between /123 and /?123?

    - by BoltClock
    I've noticed that some sites (including http://jobs.stackoverflow.com) have query strings that look like this: http://somewebapp.example/?123 as compared to: http://somewebapp.example/123 or http://somewebapp.example/id/123 What are the reasons that developers choose to implement their web apps' URLs using the first example instead of the second and third examples? And as a bonus, how would one implement the first example in PHP, given that 123 is the primary key of some row in a database table? (I just need to know how to retrieve 123 from the URL; I already know how to query the database for a primary key of 123.)

    Read the article

< Previous Page | 170 171 172 173 174 175 176 177 178 179 180 181  | Next Page >