Search Results

Search found 42685 results on 1708 pages for 'page speed'.

Page 706/1708 | < Previous Page | 702 703 704 705 706 707 708 709 710 711 712 713  | Next Page >

  • How would you measure conversion for an iPhone App Download?

    - by Eran Kampf
    I want to test how users convert from my web page to downloading my iPhone app. Right now the best funnel I figured out was to go through a page that pings Google Analytics and then redirects to an iTunes link. But this funnel only measures conversion for users who got redirected to iTunes and not if they actually downloaded the app once they saw it in iTunes. Anyone knows of a way to measure conversion to the actual download? (I know this is not a coding question but still a problem a app developers would encounter trying to market their app)

    Read the article

  • htaccess: how to prevent infinite subdirectories?

    - by Vincent Isles
    My .htaccess at the root directory contains this: Options -MultiViews +FollowSymlinks -Indexes ErrorDocument 404 /404.htm RewriteEngine On RewriteBase / followed by a bunch of RewriteRules. The .htaccess at the /forum subdirectory contains this: RewriteEngine On RewriteBase /forum RewriteRule ^index.html$ index.php [L,NE] RewriteRule ^(.*)-t-([0-9]+).html(.*)$ showthread.php?tid=$2$3 [QSA,L] RewriteRule ^(.*)-t-([0-9]+)-([0-9]+).html$ showthread.php?tid=$2&page=$3 [QSA,L] followed by other rules mapping SEO-friendly URLs to the true URLs. mydomain.com/forum/a-thread-t-1.html returns the page showthread.php?tid=1 but so does mydomain.com/forum/forum/a-thread-t-1.html, mydomain.com/forum/forum/forum/a-thread-t-1.html and so on. I don't want this behavior - I want pages accessed as /forum/forum/* to return a 404. Any hint on where I had been wrong?

    Read the article

  • Send data to webserver from C#, what's the most efficient way?

    - by Brian
    I am sending gps coordinates from a windows mobile phone to a webserver using a basic program I wrote in C#. The problem is the data plan on the phone only allows 4 MB per month. I was planning on updating the location every 10 seconds. Currently I am just creating a webrequest every 10 seconds to a php page on the server and the coordinates are passed over in the url, the php page saves them to the database. This generates about 1K of data per request, at this rate I will hit my data limit in less than a day. Is there a more efficient way to do this?

    Read the article

  • Display loading while webservice is running

    - by poseid
    On the initial request to a page, I trigger a webservice and want to display a loading message. When the page has finished loading, I want to remove the waiting message from screen, but display it again when a change event in a checkbox triggers an Ajax call with another loading message. This code is not behaving to the given spec: $(document).ready( $(function() { $('#loadingDiv').hide(); $('#productsDropDown') .change(function() { var prodValue = $(this).val(); $('#proddate').load('getpdate.php', {prod: prodValue }); }); $('#loadingDiv') .ajaxStart(function() { $(this).show(); }) .ajaxStop(function() { $(this).hide(); }) ; }); ); what am I doing wrong?

    Read the article

  • Default Values Specflow Step Definitions

    - by Gavin Osborn
    I'm starting out in the world of SpecFlow and I have come across my first problem. In terms of keeping my code DRY I'd like to do the following: Have two scenarios: Given I am on a product page And myfield equals todays date Then... Given I am on a product page And myfield equals todays date plus 4 days Then... I was hoping to use the following Step Definition to cover both variants of my And clause: [Given(@"myfield equals todays date(?: (plus|minus) (\d+) days)?")] public void MyfieldEqualsTodaysDate(string direction, int? days) { //do stuff } However I keep getting exceptions when SpecFlow tries to parse the int? param. I've checked the regular expression and it definitely parses the scenario as expected. I'm aware that I could so something as crude as method overloading etc, I was just wondering if SpecFlow supported the idea of default parameter values, or indeed another way to achieve the same effect. Many Thanks

    Read the article

  • Saving small text files is slow over Win Server 2008 R2 VPN

    - by Buckers
    We have a VPN connection to our Windows Server 2008 R2 machine, and the connection works fine. Large files go back and forth fairly quickly, but we use the connection mainly for working on small text files (.aspx, .asp, .php etc). What we find very annoying is that even the smallest of files, there is a noticeable delay of between 2-5 seconds when saving any changes. As we often make changes to code and are constantly saving, this is becoming a problem. Is there anything that might be causing this delay? Or is there anything we can do to speed it up? The connection is definitely not the issue as we have a constant 5Mb upload from our server, and 20Mb+ down on the remote machines. Thanks, Chris.

    Read the article

  • Using Selenium to Determining The Visibility of Elements for Print media

    - by Tom Howard
    I would like to determine if particular elements on a page are visible when printed as controlled by CSS @media rules. Is there a way to do this with Selenium? I know there is the isDisplayed method, which takes the CSS into account, but there is nothing I can find to tell Selenium which media type to apply. Is there a way to do this? Or is there another way to test web pages to make sure the elements you want are printed (and those you don't aren't)? Update: For clarity, there are no plans to have a javascript print button. The users will print using the normal print functionality of the browser (Chrome, FF and IE). @media css rules will be used to control what is shown and hidden. I would like Selenium to pretend it is a printer instead of a screen, so I can test if certain elements will be visible in what would be the printed version of the page.

    Read the article

  • How can I figure out which PHP extensions aren't being used?

    - by Tom Marthenal
    I manage a server (running Ubuntu) which hosts our client's sites with a few dozen different PHP-based websites, mostly small sites but also some installations of CMSes and forums. I used the get_loaded_extensions() method to see what extensions I have loaded. To help streamline the server (remove unnecessary extensions to make upgrading easier and marginally improve speed), I'd like to remove extensions that aren't being used by any of the sites. I currently have 54 different extensions loaded. I can easily eliminate some of these from the list which I know are used, but others I am less sure about. Is there some way that I can see extensions which have not been used recently?

    Read the article

  • How do I submit really big amounts of data to a form?

    - by William Calleja
    I have an HTML from that's posting a really big amount of data which is eventually being saved into an SQL Server 2005, the form is as follows: <form name="frmForm" method="post" action="saveData.aspx"> the target page takes the content of a control within the form and saves it to the database through a normal SQL insert statement. However only a portion of the data is being saved. The field in the database is an ntext. Should I use a different field? Or is something happening while I'm transferring from one page to another? Or even still there's something happening when I'm sending the really big SQL statement through c# in saveData.aspx?

    Read the article

  • SQL Server 2005 - Understanding ouput of DBCC SHOWCONTIG

    - by user169743
    I'm seeing some slow performance on a SQL Server 2005 database. I've been doing some research regarding SQL Server performance but I'm having difficulty fully understanding the output of SHOWCONTIG and would be very grateful if someone could have a look and offer some suggestions to improve performance. TABLE level scan performed. Pages Scanned................................: 19348 Extents Scanned..............................: 2427 Extent Switches..............................: 3829 Avg. Pages per Extent........................: 8.0 Scan Density [Best Count:Actual Count].......: 63.16% [2419:3830] Logical Scan Fragmentation ..................: 8.40% Extent Scan Fragmentation ...................: 35.15% Avg. Bytes Free per Page.....................: 938.1 Avg. Page Density (full).....................: 88.41%

    Read the article

  • Sphinx: change column set for searching in runtime

    - by Fedyashev Nikita
    I use Ultrasphinx gem plugin as a wrapper for accessing Sphinx search daemon. My model declaration: class Foo < ActiveRecord::Base is_indexed :fields => ['content', 'private_notes', 'user_id'] Client code: filters = {} if type == "private" # search only in `content` column filters['user_id'] = current_user.id else # search in `content` and `private_notes` columns end results = Ultrasphinx::Search.new(:query => params[:query], :per_page => 20, :page => params[:page] || 1, :filters => filters) The problem I have now with Ultrasphinx gem(or Sphinx, in general?) is that it does not allow me to change set of fields where to look for matches IN RUNTIME How can I solve this problem?

    Read the article

  • gcalFeed is not a function?

    - by fossildesigns
    I set up a calendar for the department at the university where I work (http://www.uwsp.edu/conted/calendar/default.aspx), this works just great. I then turned around and uploaded all the jquery files and fullcalendar.css to my development server (http://fossilfoundry.com/test/test.htm). When the page loads I get a JavaScript Error stating that $.fullCalendar.gcalFeed is not a function. All the files are local to the .htm page. I've checked everything I can think of. Any ideas why this won't run?

    Read the article

  • IIS Not working

    - by 3bd
    I have a web site that built on Visual studio 2008 and i need to run it from my computer (Win 7 Ultimate) as a server I tried to publish it to IIS and this is simply not working and i have the flowing error : Error Summary HTTP Error 500.19 - Internal Server Error The requested page cannot be accessed because the related configuration data for the page is invalid. Config Error This configuration section cannot be used at this path. This happens when the section is locked at a parent level. Locking is either by default (overrideModeDefault="Deny"), or set explicitly by a location tag with overrideMode="Deny" or the legacy allowOverride="false". any one can help?

    Read the article

  • Using sessions & session variables in a PHP Login Script

    - by user1325291
    I have just finished creating an entire login and register systsem in PHP, but my problem is I haven't used any sessions yet. (Like, there's not the word "session" in the entire script). I'm kind of a newbie in PHP and I've never used sessions before. What I want to do it, after the user registers and fills out the login form, they will still stay on the same page. So, like one half of the page will be if the session is logged_in and the other part will be else (the user is not logged in so display the login form). I also want to create session variables which will hold the username and password given at login. Thanks in advance! BTW this is my website - CP Cheats

    Read the article

  • JavaScript and PHP - A little confused on how to work them together.

    - by Doug
    So right now, I'm just using a basic form to check a password. I want it to check the password and basically remain on page.html so I can use JavaScript to alert incorrect password or something. I'm not really sure how to do that. It seems it would bring me to check.php. I'm not too sure on the whole process, any help appreciated! Thanks! Page.html <form action="check.php" method="post"> <input type="password" name="password" /> <input type="submit" value="Submit" /> </form> check.php <?php $password = $_POST['password']; if ( $password != "testing" ) { die(); } ?>

    Read the article

  • Rails routing issue

    - by Brob
    Hi I've just started learning rails and been going through some tutorials from learning rails (http://www.buildingwebapps.com/podcasts/79335-putting-the-page-contents-into-the) The tutorials appear to be Rails v2 and I'm on Rails v3 There is a line in the routes file which appears to be causing a problem map.view_page ':name', :controller => 'viewer', :action => 'show' The line works for the front end view i.e. my viewer controller but not for the back end I get the error NoMethodError in Viewer#show I think this has something to do with the view I am using and the line <%= @page.body %> I know it's difficult without the full code but if anyone can help that would be awesome

    Read the article

  • Best method of Zend Framework caching

    - by iamthejeff
    I have a blog built using Zend Framework, which I realize might be a bit overkill for a blog alone, but I am planning on adding other features in the future. Nevertheless, I've noticed pages could be a little speedier. I've done a basic caching method that basically captures everything in index.php (Core frontend and File backend), which works great, but unfortunately it also prevents dynamic page contents from updating (messages like "this was posted 5 minutes ago", etc) until the cache period expires. So my question is what would be the best method of caching to improve performance? I am doing fairly basic queries which are mostly simple selects, not many joins or anything fancy (using Zend_Db_Table), and even on a small database page loads are a little sluggish. Is it worth it to cache queries or should I focus my time elsewhere?

    Read the article

  • What are those CPU control panel softwares called?

    - by Tim
    My computer is running an AMD Phenom II quad core CPU but it didn't come with the piece of software that I see people have in a lot of youtube videos. The software is kind of like the ATI catalyst control center but for the AMD processor instead. It shows things like current core temperature, clock speed, etc and I am not sure but maybe it also allows the user to make changes to those things from the control center as well. I am having a hard time finding the download for this online, especially since I don't even have the name for it.

    Read the article

  • Drupal rendering incomplete views

    - by Paul
    I got tapped to do some quick maintenance on a recently migrated Drupal site. I'm pretty new to Drupal, so hopefully the problem is something that more experienced guys will figure out quick. Behavior is as follows. The public content works fine as far as I can tell. When I go to login, the login form renders correctly, but when I post the form w/ my credentials, I get back a blank page (not a 404; looks like a 200 to the /user URL, but all the gets rendered is empty Head and body tags). If i refresh the page, I get the content of my profile view, but none of the site chrome or CSS. Note that this is not an issue on the site it was migrated from, so it seems like something wasn't copied over correctly. The site's not public, so I can't provide a URL, sorry!

    Read the article

  • Cookie Blocked popup in IE

    - by Shyju
    I have a website hosted in a server named msa_inst_server. In my login page ,after user enter the username,i will check whether cookies are available or not,But when i access the login page in IE 6.0 in one machine in the local network,I am getting a pop up saying that I have added the site to trusted zone,changed the settings to allow all cookies.But no way to get rid of this problem. Any thoughts.... ? Does the underscore in the server name matters ? Thanks in advance

    Read the article

  • How to enable indexing of pages with dynamic data?

    - by mithunb
    I have a site that has certain urls that point to pages with permanent data and others that point to dynamic web pages. Google indexes both these regularly. By the time a user finds one of the dynamic content urls, the data on the page has already changed and the user does not find what he was looking for. Further, the dynamic url pages contains links to the permanent urls (which I want Google or any crawler to index). Google crawler controls (webmaster tools) cannot be made to read urls from a page but not index them. Solutions? crawling strategies *system architecture*.

    Read the article

  • Installing Windows on multiple computers

    - by Rob
    At our work we've decided to buy SSDs to speed up our computers. We'd also like to upgrade from Windows 7 to Windows 8.1 (read clean install) I'm talking about 10 computers which going to have a fully clean installation. Is there any trick I can use to avoid installing Windows 10 times? The computers (most of them) are different in hardware, so I think duping is no good option. What can I do? I'd like to spend as less time as possible because all computers are going to need SQL Management Studio and Visual Studio 2012/2013. Thanks!

    Read the article

  • Replace the surround of an html element with another document

    - by KvanteTore
    I have an html page with (among other things) a Unity3D window. I would like to replace everything on the page without causing the Unity window to reload. I have tried the following jquery-tastic function replaceSurround(keepElem, newElem) { keepElem.siblings().remove(); keepElem.prepend(newElem.prevAll()); keepElem.append(newElem.nextAll()); var keepParent = keepElem.parent(); var newParent = newElem.parent(); if (keepParent && newParent) { replaceSurround(keepParent, newParent); } } where keepElem is an element in the original document and newElem is the corresponding element in the new document, but it did not work very well.

    Read the article

  • MS SQL 2005 - Understanding ouput of DBCC SHOWCONTIG

    - by user169743
    I'm seeing some slow performance on a MS SQL 2005 database. I've been doing some research regarding MS SQL performance but I'm having difficulty fully understanding the output of SHOWCONTIG and would be very grateful if someone could have a look and offer some suggestions to improve performance. TABLE level scan performed. Pages Scanned................................: 19348 Extents Scanned..............................: 2427 Extent Switches..............................: 3829 Avg. Pages per Extent........................: 8.0 Scan Density [Best Count:Actual Count].......: 63.16% [2419:3830] Logical Scan Fragmentation ..................: 8.40% Extent Scan Fragmentation ...................: 35.15% Avg. Bytes Free per Page.....................: 938.1 Avg. Page Density (full).....................: 88.41%

    Read the article

< Previous Page | 702 703 704 705 706 707 708 709 710 711 712 713  | Next Page >