Search Results

Search found 16097 results on 644 pages for 'dynamic pages'.

Page 212/644 | < Previous Page | 208 209 210 211 212 213 214 215 216 217 218 219  | Next Page >

  • Win-XP Browsers Hang on page load - (waiting for...)

    - by CHarmon
    Hello, I’m having problems with my browsers hanging on loading pages on my desktop machine. I’m using Windows XP Pro with SP3 and fully updated except for IE 8. All three of my browsers, IE 7, Chrome and Firefox are having the same problems. Pages are not being loaded and are hanging on “waiting for …”. The browsers are waiting for the page being loaded or ad servers. Sometimes a page will load but the loading graphic continues to be displayed as if the page were still loading when the page appears to be fully loaded. The problem is bad enough that I can’t really use any of my browsers. I can eventually get most pages to load by stopping and restarting the page load. I have DSL modem with a wireless router and I have been able to eliminate the modem and router from being the source of my problem. My laptop doesn’t have any problems even when hardwired to the router and with the wireless connection disabled. I deleted the NIC and let XP re-install. Also tried a different network cable. Tried the same router port used in the laptop test. One clue that may be important is that I can’t connect to my router using the desktop machine…the page hangs while trying to connect. I can ping the router and I can quickly connect to the router using the laptop. I also can’t use the Windows update process – the page never fully loads. The problem affects other user accounts and even happens in safe mode. I am convinced the problem is with part of the O/S…some layer able to affect all of the browsers. The purpose of this post is to see if anyone has some ideas before I do a XP repair. I have done quite a bit of trouble-shooting: Ran a full anti-virus scan with AVG – no problems. Ran full scans with Spybot, MalwareBytes and Sophos anti-rootkit – no problems. Ran Chkdsk with both options checked. Ran Disk Clean up Defragged RE-installed IE7 Cleared all the browser caches Ran Ccleaner (registry tool) Ran HijackThis – nothing unusual (problem happens in safe mode too) Ran Process Explorer – no unusual processes Used System Restore and fell back several days – no change in the problem Booted to last known good configuration – no change in the problem Ran MicrosoftFixit50199.msi – no change in the problem Any ideas or suggestions would be appreciated…I’m not looking forward to doing a repair on XP. Thanks in advance for any help.

    Read the article

  • GhostScript noob help - Breaking a multipage PDF file into many single page PS or EPS files.

    - by godzilla_g
    Hi, I'm trying to do the following with ghostscript: Turn one multipage PDF file (about 3,000 pages, 200mb file) into: One file per page of the PDF, and convert each (page/file) to EPS or PS (post script(preferably)). Example: hello.pdf (10 pages) would produce: hello1.ps (page 1 out of 10) hello2.ps hello3.ps ... hello10.ps How can I do this? I've been trying for 4 days, and can't figure it out. I have a script I've tried(won't work): Note: Windows(7) user here. gs -sDEVICE=epswrite -o documentname-%.eps documentname.pdf I also don't know how to navigate to the directory where my file resides (cannot figure that out, too). If you can, please show me how. A big thank you.

    Read the article

  • Problem with internet connection

    - by vijay.shad
    Hi all, I am working on windows vista. I have got two internet connections, one is wifi connection with high speed. And other is mobile network connection. There is a very strange problem I am getting. When I connect to wi-fi connection I an not able to serf internet(actually not all the sites). I am able to search on Google but when i click on any link in the search list it does not open. But I am able to serf all the pages in google.com domain and also all the pages in stackoverflow domain. But i am not able to go to page http://repo1.maven.org/maven2 But When I am connected by my mobile network. I am able to serf any site. Can you please tell me what might me the problem with my settings.

    Read the article

  • Using Apache Environment Variables to set custom ErrorDocument

    - by Tad
    I've got a set of RewriteCond rules that test for various mobile devices and then set environment variables like "env=device:.iphone" or "env=device:.smartphone" if the useragent matches an iPhone or Android device. I'm trying to now redirect the user to custom-styled 404/500 server error pages for each device, by way of the error pages. Ideally I'd like to be able to test for a variable being there, and then write in a custom ErrorDocument string. But an apache doesn't seem to work in this case. Any ideas how I can construct if/else tests in an apache conf file for environment vars?

    Read the article

  • Google "not selecting" many of the links on my site

    - by Loki
    Since a few days I noticed that Google wasn't indexing any of the pages on my site anymore. When I checked the indexingstatus-page I noticed the following: As you can see the "not selected"-line is going thru the roof! The information on google's help pages is very limited: https://support.google.com/webmasters/bin/answer.py?hl=nl&answer=139066 I run a free downloads-site which is automatically updated. I don't have duplicate content (except that sometimes descriptions of software is similar over an older version of the software) and my URL's are all forwarded to the www.-variant from within Wordpress. So the canonical-part that Google mentions in their Help-file, isn't the problem. Any ideas what could be causing this and how to solve it?

    Read the article

  • Recommend a Software for documenting work/research

    - by Rogue
    Need a software for documenting my work & research. I was considering to use blog engine but then I don't want to be playing around with backing up the databases and the matter separately. The specifications for the software would be: 1) To catalogue work by categories, tags and date. 2) Easy editing and drawing features 3) Easily incorporate saved html files into matter I'm writing, with the formatting and pictures. (probably use the mht format, don't want to be wasting timeformatting html pages) Would like the html pages to be directly displayed in the software and not opened on demand Software should be for Windows. Any recommendations? 4) Language tools could be an addition but only an optional feature

    Read the article

  • Impersonation on IIS 7.0 passes the machine credentials for Crystal Reports

    - by pknox
    On a 32-bit Windows 2008 server running the Donor2 Application in the Classic .NET Managed Pipeline mode, configured for Windows Integrated Authentication and Impersonation, all of the .NET pages are passing the authenticated user’s credentials [DomainName\UserName]. This is the correct, expected behavior. The Crystal Reports pages, instead of passing the authenticated user’s credentials, are passing the IIS Server’s credentials [DomainName\MachineName$]. One of the very frustrating aspects of this situation is that I have another server which, as far as I can tell, is configured identically. That server, when loading Crystal Reports, is passing the authenticated user’s credentials [DomainName\UserName] as expected. I have obviously missed something, but I have no idea what it could be.

    Read the article

  • Linux (non-transparent) per-process hugepage accounting

    - by Dan Pritts
    I've recently converted some java apps to run with linux manually-configured hugepages. I've got about 10 tomcats running on a system and I am interested in knowing how much memory each one is using. I can get summary information out of /proc/meminfo as described in Linux Huge Pages Usage Accounting. But I can't find any tools that tell me about the actual per-process hugepage usage. I poked around in /proc/pid/numa_stat and found some interesting information that led me to this grossity: function pshugepage () { HUGEPAGECOUNT=0 for num in `grep 'anon_hugepage.*dirty=' /proc/$@/numa_maps | awk '{print $6}' | sed 's/dirty=//'` ; do HUGEPAGECOUNT=$((HUGEPAGECOUNT+num)) done echo process $@ using $HUGEPAGECOUNT huge pages } The numbers it gives me are plausible, but i'm far from confident this method is correct. Environment is a quad-CPU dell, 64GB ram, RHEL6.3, oracle jdk 1.7.x (current as of 20130728)

    Read the article

  • latex page number position

    - by anton
    I am working on a long document in latex with documentclass book. I need the page number to always be in the upper right corner of each page, even if that page is the first page of a chapter (right now on 1st pages of chatpers, the page number is bottom-centered, on all other pages it's top-right). I control the position of the page number with fancyheader: \usepackage{fancyhdr} \pagestyle{fancy} \lhead{} \chead{} \rhead{\thepage} \lfoot{} \cfoot{} \rfoot{} \renewcommand{\headrulewidth}{0pt} Also, I don;t know if the problem is realted but my chapters do not start from the top of the page. There is a white area, then comes chapter X, then a newline with the chapter line. What I also want is the chapter to start from the top of the page. The main question here is how I can get the page number to always appear in the upper right corner, I mention the thing with the chapter title position only in case that might be related. Thanks.

    Read the article

  • Is Software Raid1 Using mdadm with a Local Hard Disk and GNDB Possible?

    - by Travis
    I have multiple webservers which use many small files to created dynamic web pages. Caching the web pages isn't an option. The webserver also performs writes so I need a synchronous filesystem. I'm looking to maximise performance as it's my understanding that small files is the weakness (to varying degreess) of a cluster filesystem over ethernet. Currently I'm using Centos 5.5, 64 bit. Since it's only about 300MB of data, I'm looking at mdadm using RAID-1 with the GNBD and a local hard disk using the "--write-mostly" option so the reads are done using the local hard disk. Is this possible? If so, is there any advantage to making it a tmpfs disk instead of a local hard disk? Or will the files on the local hard disk just get cached in RAM anyway so I won't see a performance gain by using tmpfs, assuming there's enough RAM available?

    Read the article

  • What's up with tab order on my Mac?

    - by biged781
    So, I just got my first Mac. It is slick, and I feel like I don't know how to do anything, but overall it is a great machine. However, I am becoming frustrated with the tab order in most web pages. For example, this site. If I am composing a comment and press tab, focus is set to the address bar. I would like the focus to shift to the button next to the text area, but no luck. Also, I cannot seem to tab into combo boxes in form pages. What is going on here exactly? This happens in FireFox as well as Safari. I don't get why the tab order of a page would not be respected. Any help is appreciated.

    Read the article

  • Error opening microsoft.com with Google Chrome

    - by Ilnur
    Hi! I need help with my browsers. In my OS (windows 7 basic) installed 2 browsers: IE 8 and Google Chrome. When I trying to open pages, which associated with microsoft.com (like ms live and others), my browser crashes. IE 8 crashes too. But Opera (I've installed it later) works fine with that pages. P.S. I tried to search the trojan software and viruses. But my PC is clean.

    Read the article

  • Troubleshooting a slow database server with no load

    - by user1721724
    I'm getting ready to soft launch my website and I've run into some problems with what I think is being caused by my MySQL database running on Fedora. All websites run fine, just as I'd expect, but any pages that establish a database connection hang until the connection is established, and then bang, the site loads as it should. Ex. My landing page (http://www.thrusong.com) doesn't make a database connection and loads quickly. User profile pages (http://www.thrusong.com/john) make a database connection and load slowly, even though most of the data comes from memcached and the database currently has no load on it. This problem just came up yesterday when my router died and I began using my Pace 2Wire modem with built-in router. Before, my old router was set to handle everything. My ISP says the settings in the modem are correct. Any ideas? Thanks in advance.

    Read the article

  • Using my System as the server- need advice

    - by Ashwin
    We are deploying my web application in my system in jboss. And I am planning to use my system as the server as well. There are no html pages or jsp pages deployed. The client requests for resources and the server provides the resources in the form of objects. We are also using databases(Postgresql-15 tables) as the database server(this will also reside in my machine). My system configuration is Windows Vista, 2.3 GHZ, 4GB Ram, 32 bit. There can be many requests coming in at the same time. So is this configuration enough? Should we go with a different Operating System like Windows Server? I have never used Windows Server OS. How will it be different from other windows operating systems? Or if you feel please some other OS will be good in this situation, please suggest.

    Read the article

  • Broadband Traffic Question

    - by rutherford
    I have a broadband ADSL line with plus.net in the UK. Having checked the modem there is no firewall or any weird features enabled. But since I arrived at the apartment (the broadband already being installed), I cannot log into Twitter nor update any of my wordpress blogs (I can browse them and log in, but cannot save any edits or new posts). It only seems to affect these two sites in their unique ways. If I take the netbook I use in this place out to say a McDonalds or some other wifi access point then these sites work fine again. Anyone know what could possibly be preventing access of the pages in question? The only thing common to these pages are the POST response they are expecting. But POST form submission works fine on other sites...

    Read the article

  • Google Chrome gets really slow on 10+ open tabs

    - by Anton
    For some time already I face a problem with Google Chrome. I really love this browser, but on Windows 7 on a pretty decent machine (i5, 4GB RAM) it gets REALLY slow when I open for instance 10 techcrunch.com pages. Once I do that it becomes virtually difficult to scroll through pages and the general responsiveness of the browser gets down. And if I open 20+ or 30+ tabs there is a good chance all of them will crash. Does anyone got an idea? This happens to me on several PCs with Windows 7 64bit. At 10 tabs there is 600-700MB memory used by Chrome. Two systems have the issue are both laptops with integrated graphics. One by Intel, the other an nVidia GeForce 310M.

    Read the article

  • I want to remove all images from my word document without losing the data formatting

    - by Chella
    Can any body help me, in this regard. I just want to remove all the images in my word document. I have about 365 pages document, in which there are images in almost all pages. I can not delete one by one. I can save it in .txt format but I will lose the formatting of the data. Is there any other way to get this. any help would be greatly appreciated. Thanks in advance. Let me know if you people need more clarification on this.

    Read the article

  • What is the Optimal Server Configuration for Split-Path Testing?

    - by doug
    I am far from an expert on Apache or any server for that matter, so i apologize if this question is poorly worded, which it likely is. We have always relied on a vendor for split-path testing (aka "AB Testing"). If you're not familiar with that term, it's a form of marketing research in which you slightly modify one of your web pages (usually one nearest the point of conversion), say for instance, by changing the position of the "Buy Now" button or its color/contrast/texture, then serving one of those two pages to a given user based on random selection. By doing split-path testing ourselves, I suspect we can do it far more cheaply and increase cycle times as well. What is the optimal set-up for these tests? "Optimal" is based on the following criteria: how quickly/easily new tests can be set-up and put online; and minimal disruption to overall site performance

    Read the article

  • Is "Cleversoft PC Tuneup Maestro" malware?

    - by Barend
    I found a desktop icon for something called PC TuneUp Maestro on my parents' laptop. Googling it yields pages upon pages of freeware download sites, but I couldn't find anything legitimately confirming or clearing it of malware status. I distrust these programs by default, but Microsoft Security Essentials seems to think it's legit. Does anyone know what this program is and if it's up to anything surreptious? Can I uninstall it or do I write off the entire OS install as tainted? The system hosts file is intact (no funny unicode lookalikes either), so that red flag, at least, hasn't been raised.

    Read the article

  • How Do I Delete Misnamed Tagged Directory Already Committed?

    - by Teno
    I'm new to using the SVN system and having hard time figuring out how to delete files uploaded mistakenly. What I've done: Committed the trunk folder with right clicking on "SVN Commit" Right clicked and choose "TortoiseSVN" - "Branch/Tag" In the section of "To path:" in the "Branch/Tag - Tortoise" window, I typed /*mydirectory*/tags/*1.0.11* where 1.0.11 was supposed to be 1.0.1.1 After realizing 1.0.11 was a mistake, to remove the directory, I right clicked on the 1.0.11 folder in Windows and selected "TortoiseSVN" - "Delete" It deleted the folder in Windows but does not delete the folder in the remote server. According to this page,http://stackoverflow.com/questions/2092344/how-do-i-delete-a-wrongly-tagged-directory-in-svn, a command can be used and I tried to type svn in the command prompt window but it gives svn is not recognized as an internal or external command. This should be a very basic question but I could not find relevant pages. Some pages suggest to use revert but I've already committed 1.0.1.1 so I'm afraid doing revert causes the newest one to be deleted. Thanks for your information.

    Read the article

  • Software RAID 1 broken, how do I fix this?

    - by Edward
    I'm running CentOS 6 x86_64. There is a software RAID 1 being used on the two internal 80GB drives. I got the following e-mail sent to me: A DegradedArray event had been detected on md device /dev/md1. Faithfully yours, etc. P.S. The /proc/mdstat file currently contains the following: Personalities : [raid1] md0 : active raid1 sda1[0] 511988 blocks super 1.0 [2/1] [U_] md1 : active raid1 sda2[0] 8190968 blocks super 1.1 [2/1] [U_] bitmap: 1/1 pages [4KB], 65536KB chunk md4 : active raid1 sdc1[0] sdb1[1] 1953512400 blocks super 1.2 [2/2] [UU] md3 : active raid1 sdd5[1] sda5[0] 61224892 blocks super 1.1 [2/2] [UU] bitmap: 1/1 pages [4KB], 65536KB chunk md2 : active raid1 sdd3[1] sda3[0] 8190968 blocks super 1.1 [2/2] [UU] unused devices: <none> The system appears to have booted fine and is working. The two drives' content did not change at all. I only removed and reinstalled them while I was booted on the CentOS Live DVD. How do I get the array working again?

    Read the article

  • How to configure and access IIS in windows 7

    - by user28233
    I've installed IIS on Windows 7(Ultimate) by going to control paneladd/remove features/ then checking the IIS. How do I access something like a preference editor like in wamp server. In wamp server I go to http://localhost/phpmyadmin. And accessing the pages I made in http://localhost/pages/ What about in IIS? How do I access something like phpmyadmin. And how do I access the webpages that I made. This is only for educational purposes. And I do not need a site to be accessible on the net. And I already have something like this: What do I do with it. Please help,thanks

    Read the article

  • How can I run my program on a large number of computers? [closed]

    - by zenpoy
    I'm looking for a (preferably free) service for running an executable I wrote? It's not malicious, it's not a virus, it's not scam, and if this is really important I can upload the python source code instead. I wrote a small crawler to gather information regarding the style of web pages for my MA project, and I need a lot more data. EDIT Here is more information on my problem and how I approach on solving it, and where I'm stuck. As part of my research I'm trying to classify text based on it's style (font-family for now), my data is based web pages, so I wrote a client/server application - the client is a crawler that gathers this data and send it to the server. The problem is that like 99% of the internet is Arial, Verdana and Helvetica - other fonts are far more rare, so I need to spend very long time to gather enough data regarding these fonts. Hope this explains it.

    Read the article

  • What is Paging in memory management?

    - by Fasih Khatib
    I was just reading Operating System Principles by Silberschatz et al when I came across paging in memory management.I'm slightly confused about it. It states that Physical Memory(I assume it's RAM) is divided into frames, and logical memory is divided into pages. CPU generates logical addresses containing page number and an offset. This page number is used to retrieve the frame number from a page table which gives the base address so the physical address is calculated as base+offset. My question is: is the page table maintained for every process? I logically think that the answer would be yes as every process will need to map its own pages to frames. I may be wrong. Please clarify. Also: paging and segmentation(where 'holes' are created in memory) are two totally different techniques that are not used in combination. Correct?

    Read the article

  • IIS + PHP + Page with lots of images = Intermittent 403 errors

    - by samJL
    I am using an up-to-date Server 2008 R2 Datacenter, running IIS 7.5 and PHP 5.3.6/FastCGI On PHP pages with lots of images (60+), some of the images fail to load It is not always the same images-- on each page refresh an image that worked previously may not load, while an image that did not now does Looking at the Net tab in Firebug reveals that the failing image requests are 403 errors All of the images are located on the server in question, and the images directory has the correct permissions I believe this problem is the result of a limit on requests All of my attempts at researching this problem point to maxConnections setting in IIS, yet mine is set at the highest/default of 4294967295 (maxBandwidth too) I am also running a ColdFusion site on the same IIS installation, and it does not suffer from 403's on pages with lots of images I am left thinking that there is another connection limit (in PHP or FastCGI?) overriding the IIS connection limit I don't see anything that looks like a request limit in the php.ini, what am I missing? Any help would be appreciated, thank you

    Read the article

< Previous Page | 208 209 210 211 212 213 214 215 216 217 218 219  | Next Page >