Search Results

Search found 9935 results on 398 pages for 'pages'.

Page 73/398 | < Previous Page | 69 70 71 72 73 74 75 76 77 78 79 80  | Next Page >

  • SQLS Timeouts - High Reads in Profiler

    - by lb01
    I've audited a SQLS2008 server with Profiler for one day.. the overhead didn't seem to trouble this new client my company has. They are using a legacy VB6 application as a front-end. They're experiencing timeouts once SQLS RAM usage is high. The server is currently running x64 sqls2008 on a VM with nearly 9 GB of RAM. SQL Server's 'max server memory option' is currently set to 6GB. I've put the results of the trace in a table and queried them using this query. SELECT TextData, ApplicationName, Reads FROM [TraceWednesday] WHERE textdata is not null and EventClass = 12 GROUP BY TextData, ApplicationName, Reads ORDER BY Reads DESC As I expected, some values are very high. Top Reads, in pages. 2504188 1965910 1445636 1252433 1239108 1210153 1088580 1072725 Am I correct in thinking that the top one (2504188 pages) is 20033504 KB, which then is roughly ~20'000 MB, 20GB? These queries are often executed and can take quite some time to run. Eventually RAM is used up because of the cache fattening, and timeouts occur once SQL cannot 'splash' pages in the buffer pool as much. Costs go up. Am I correct in my understanding? I've read that I should tune the associated T-SQL and create appropriate indices. Obviously cutting down the I/O would make SQL Server use less RAM. OR, maybe it might just slow down the process of chewing up the whole RAM. If a lot less pages are read, maybe it'll all run much better even when usage is high? (less time swapping, etc.) Currently, our only option is to restart SQL once a week when RAM usage is high, suddenly the timeouts disappear. SQL breathes again. I'm sure lots of DBAs have been in this situation.. I'm asking before I start digging out all of the bad T-SQL and put indices here and there, is there is something else I can do? Any advice except from what I know (not much yet..) Much appreciated. Leo.

    Read the article

  • If a server is performing slowly, will ping still appear normal?

    - by blarg
    I'm trying to troubleshoot a periodic slow application problem without direct access to the network. There is a server on one network providing webpages to a VLAN on a different network via a DMZ. I'm told that when the pages were loading slowly the user pinged the server and saw an extremely slow response time. My question is, as ping is minimal in network load and a different protocol from the pages, would a slow ping be caused by a slow server, or would it indicate a slow network?

    Read the article

  • Why is my concurrency capacity so low for my web app on a LAMP EC2 instance?

    - by AMF
    I come from a web developer background and have been humming along building my PHP app, using the CakePHP framework. The problem arose when I began the ab (Apache Bench) testing on the Amazon EC2 instance in which the app resides. I'm getting pretty horrendous average page load times, even though I'm running a c1.medium instance (2 cores, 2GB RAM), and I think I'm doing everything right. I would run: ab -n 200 -c 20 http://localhost/heavy-but-view-cached-page.php Here are the results: Concurrency Level: 20 Time taken for tests: 48.197 seconds Complete requests: 200 Failed requests: 0 Write errors: 0 Total transferred: 392111200 bytes HTML transferred: 392047600 bytes Requests per second: 4.15 [#/sec] (mean) Time per request: 4819.723 [ms] (mean) Time per request: 240.986 [ms] (mean, across all concurrent requests) Transfer rate: 7944.88 [Kbytes/sec] received While the ab test is running, I run VMStat, which shows that Swap stays at 0, CPU is constantly at 80-100% (although I'm not sure I can trust this on a VM), RAM utilization ramps up to about 1.6G (leaving 400M free). Load goes up to about 8 and site slows to a crawl. Here's what I think I'm doing right on the code side: In Chrome browser uncached pages typically load in 800-1000ms, and cached pages load in 300-500ms. Not stunning, but not terrible either. Thanks to view caching, there might be at most one DB query per page-load to write session data. So we can rule out a DB bottleneck. I have APC on. I am using Memcached to serve the view cache and other site caches. xhprof code profiler shows that cached pages take up 10MB-40MB in memory and 100ms - 1000ms in wall time. Pages that would be the worst offenders would look something like this in xhprof: Total Incl. Wall Time (microsec): 330,143 microsecs Total Incl. CPU (microsecs): 320,019 microsecs Total Incl. MemUse (bytes): 36,786,192 bytes Total Incl. PeakMemUse (bytes): 46,667,008 bytes Number of Function Calls: 5,195 My Apache config: KeepAlive On MaxKeepAliveRequests 100 KeepAliveTimeout 3 <IfModule mpm_prefork_module> StartServers 5 MinSpareServers 5 MaxSpareServers 10 MaxClients 120 MaxRequestsPerChild 1000 </IfModule> Is there something wrong with the server? Some gotcha with the EC2? Or is it my code? Some obvious setting I should look into? Too many DNS lookups? What am I missing? I really want to get to 1,000 concurrency capacity, but at this rate, it ain't gonna happen.

    Read the article

  • I need to move part of my website to another web server, how can I do this and keep the same Domain

    - by hamlin11
    I need to move a section of my website to another server because it is taxing our current web server. However, I cannot afford to lose page rank on any pages within the section of the site that must be moved. Furthermore, the URL of the pages must not be changed... visitors must still see the same URL, even though it would be served up by different hardware from a different data center. How can this be accomplished? Thanks

    Read the article

  • Network Tracking

    - by jeswin14
    Hi all, I am having fifty workstations which are connected to a windows server 2003 server and the internet is shared from the server .My question is , am i able to implement a tracking mechanism without using any proxy server and track all the web pages which ever visited by an individual workstation connected to a network?If possible how to achieve that kind of tracking mechanism?Some workstations do open their web pages in InPrivate browsing , can we access those browsing history?

    Read the article

  • SQLS Timeouts - High Reads in Profiler

    - by lb01
    Hi I've audited a SQLS2008 server with Profiler for one day.. the overhead didn't seem to trouble this new client my company has. They are using a legacy VB6 application as a front-end. They're experiencing timeouts once SQLS RAM usage is high. The server is currently running x64 sqls2008 on a VM with nearly 9 GB of RAM. SQL Server's 'max server memory option' is currently set to 6GB. I've put the results of the trace in a table and queried them using this query. SELECT TextData, ApplicationName, Reads FROM [TraceWednesday] WHERE textdata is not null and EventClass = 12 GROUP BY TextData, ApplicationName, Reads ORDER BY Reads DESC As I expected, some values are very high. Top Reads, in pages. 2504188 1965910 1445636 1252433 1239108 1210153 1088580 1072725 Am I correct in thinking that the top one (2504188 pages) is 20033504 KB, which then is roughly ~20'000 MB, 20GB? These queries are often executed and can take quite some time to run. Eventually RAM is used up because of the cache fattening, and timeouts occur once SQL cannot 'splash' pages in the buffer pool as much. Costs go up. Am I correct in my understanding? I've read that I should tune the associated T-SQL and create appropriate indices. Obviously cutting down the I/O would make SQL Server use less RAM. OR, maybe it might just slow down the process of chewing up the whole RAM. If a lot less pages are read, maybe it'll all run much better even when usage is high? (less time swapping, etc.) Currently, our only option is to restart SQL once a week when RAM usage is high, suddenly the timeouts disappear. SQL breathes again. I'm sure lots of DBAs have been in this situation.. I'm asking before I start digging out all of the bad T-SQL and put indices here and there, is there is something else I can do? Any advice except from what I know (not much yet..) Much appreciated. Leo.

    Read the article

  • what is the main cause of 500 internal server error? [closed]

    - by Usman
    I want to know that I have hosted with a hosting company . My website gives "500 Internal server error many times" I have following Web server statistics :- Web Server Statistics Successful requests: 127,310 (7,504) Average successful requests per day: 814 (1,071) Successful requests for pages: 24,949 (1,309) Average successful requests for pages per day: 159 (186) Failed requests: 3,499 (58) Redirected requests: 10,091 (114) Distinct files requested: 5,791 (556) Distinct hosts served: 5,107 (330) Data transferred: 4.28 gigabytes (190.56 megabytes) Average data transferred per day: 28.03 megabytes (27.22 megabytes) Can you tell me my server condition by seeing this or i have to give another details. Thanks in advance

    Read the article

  • Nginx error page using django master page

    - by user835199
    I am using python django to develop a web app and using nginx and gunicorn as servers. I need to define nginx error page (for error codes like 500, 501 etc), but i want to keep the layout same as that in other site pages. For site pages, i use the include functionality of django but in this case, since django won't preprocess the page, i need to create a pure html page. Is there a way to reuse the master page that i created in django for creating this error page?

    Read the article

  • Network Tracking

    - by jeswin14
    Hi all, I am having fifty workstations which are connected to a windows server 2003 server and the internet is shared from the server .My question is , am i able to implement a tracking mechanism without using any proxy server and track all the web pages which ever visited by an individual workstation connected to a network?If possible how to achieve that kind of tracking mechanism?Some workstations do open their web pages in InPrivate browsing , can we access those browsing history?

    Read the article

  • Displaying Multiple Tables in a Page in Jasper Reports using ireports/Jasper Assistant

    - by Karthikeyan
    Hi all, I have a report to be printed / exported to PDF format. for that i am using jasper reports. I have 8 individual pages to be exported. so i prepared 8 separate design files. In that 6 pages are having more static text and mere direct mapping and remaining two pages consists of table like data. one page contains huge table so i filled it directly with no problem by keeping a single table in detail band. now problem is that remaining one page. that page is having more than 5 small tables (two or four column). is t possible to have more that one table in a page in jasper reports ? i am using jasper reports and for designing jasper assistant/ireports.

    Read the article

  • Why won't this Apache modrewrite RewriteRule work?

    - by Jason Rhodes
    I'm trying to get Apache mod rewrite to work on my local machine. I'm running OSX with PHP 5 and the Apache mod rewrite module is enabled. I have a directory called localhost/~Jason/hfh/admin with various PHP includes called based on a $_GET variable. I want to let users type (in theory) localhost/~Jason/hfh/admin/pages and have that URL stay in the address bar, while what gets displayed is localhost/~Jason/hfh/admin/?admin=pages So. I've created a .htaccess file that sits in the /hfh directory. Inside, I've put this mod rewrite text: RewriteEngine On RewriteRule ^admin/([^/.]+)/?$ admin/?admin=$1 [L] When I go to the browser and type localhost/~Jason/hfh/admin/pages I get a "Problem loading page" error, and Firefox says, "Oops. Firefox can't load this page for some reason." Can anyone help me figure this out? I have such a hard time with regex and mod rewrite...

    Read the article

  • JavaEE in netbeans giving BUILD FAILED error upon deployment

    - by user312402
    When I try to run my Java EE program in Netbeans consisting of servlets (java pages), JSP's, beans(java pages) and HTML pages I get this error in the output: In-place deployment at C:\Users\Derek\Documents\NetBeansProjects\EJBProject\EJBProject-war\build\web Initializing... deploy?path=C:\Users\Derek\Documents\NetBeansProjects\EJBProject\EJBProject-war\build\web&name=EJBProject-war&force=true failed on Personal GlassFish v3 Domain C:\Users\Derek\Documents\NetBeansProjects\EJBProject\EJBProject-war\nbproject\build-impl.xml:611: The module has not been deployed. BUILD FAILED (total time: 1 second) And then in the command prompt when I run asant run in the appropriate directory, I get: C:\Users\Derek\Documents\NetBeansProjects\EJBProject\nbproject\build-impl.xml:19: Class org.apache.tools.ant.taskdefs.condition.Not doesn't support the nested "antversion" element. Do you know why this would be? Why won't netbeans deploy my application so I can run and test it?

    Read the article

  • Treeview - Link Button Post Back Problem

    - by cagin
    Hi there I' m working on a web application. That has a master page and two pages. These pages under the that master page. I am trying navigate that pages with a TreeView which on the master page. When i click to treeview node i can go to page which i want but there is no postback. But if i use linkbutton postback event happen. I use a break point on master page's pageload event. When i use treeview, v.s doesnt stop on break point line but if i use link button v.s stop on that line. How can i do postback with using treeview? Thanks for your helps

    Read the article

  • JSF Managed Bean auto-create?

    - by rat
    Is it possible to have a JSF managed bean be automatically created? For example I have several session scoped beans. Sometimes it becomes necessary to access these instances in code (rather than just in JSF) this is done by: PageBean pageBean = (PageBean) FacesContext.getCurrentInstance().getExternalContext().getSessionMap().get("pages"); However if no page has already been visited which calls to '#{pages}' this resolves to null ... is there anyway to get JSF to create a bean when the scope 'begins'? So in this case ideally when a user session begins the 'pages' bean would be instantiated in the session immediately?

    Read the article

  • MembershipProvider, IPrincipal, IIdentity?

    - by MRFerocius
    Hello guys; I have a conceptual question... I am making an Intranet application (Web platform) for a company. I have a SQL Server DB with these tables: Users (userID, userName, userPass, roleID) Roles (roleID, roleName) Pages (pageID, pageURL) RolesXPages(pageID, roleID) How is the best way to create a structure to store all this information while the user navigates the site, I mean, on the thread I should be able to check his role, his pages (the ones he can access) I have been reading and there is a lot of stuff there where Im confused, I saw the MembershipProvider, IPrincipal, IIdentity, etc classes but Im not sure what should be the best one for me. Any thoughts... Thanks in advance! Edit: Everytime gets more confusing... I just want to handle those structures at runtime and be able to mantain state during page callbacks or changing pages...

    Read the article

  • Do not fetch app.manifest each time

    - by Kristof
    For creating an offline version of a bunch of linked web pages I use an app.manifest-file that lists all the web pages for offline caching. I would like it that the app.manifest file is not fetched every time when a user jumps from one web page to another. Most of the web pages will never be updated once the application is on the iPhone. Also the target audience is abroad so roaming costs could add up while users are using my "web app". Searching the internet made me think in the direction of forcing an offline mode in Mobile Safari using Javascript but I don't know if this is the right way to go or if it is even possible. Does anyone have any other and/or better ideas and suggestions on how to do this?

    Read the article

  • PHP Explode with an Unicode character as separator

    - by Young Roger
    XPDFs pdftotext converts pdf to text and outputs it at command line level. If needed it inserts PageBreaks between the pages as specified in TextOutputDev.cc: eopLen = uMap->mapUnicode(0x0c, eop, sizeof(eop)); This Unicode symbol is encoding independent, -enc ASCII7 wouldn't change it. I'm currently willing to use PHP for converting and splitting the PDF file into several TXT pages for database storage. However, the following function does work, but takes twice as long as a conversion of the whole book in one time. for($i = 1; $i <= $pages[0]; $i++) $page[$i] = shell_exec('/usr/bin/pdftotext sample.pdf -f '.$i.' -l '.$i.' -'); How am I supposed to explode(0x0c, $wholePDF) with an Unicode character as separator? Currently, page[$i] doesn't seem to retrieve those weird Unicode PageBreak characters from the shell_exec(). I tried several headers for encoding (UTF-8 especially) but it didn't work out so far.

    Read the article

  • Bulk import 301 redirects in IIS 6.0

    - by pkiddie
    Is there a mechanism, or any additional tools that I can make use of to bulk import 301 redirect rules into an IIS6.0 web application? I want to be able to 301 redirect some old .htm pages to their new home on a CMS to ensure the current search index entries resolve OK and don't 404, but there are roughly ~50 pages. The only way I've found to achieve this is to create dummy .htm files and then right click on the file-File tab-Redirect to a URL radio button within the IIS MMC snapin, which is a little laborious over the number of pages. Thanks.

    Read the article

  • Lobo Cobra HtmlPanel does not update GUI after JavaScript action

    - by Markus Lausberg
    HI, I started implementing a HTML Browser in Swing. I want to use the Lobo Cobra toolkit because this toolkit inclused a renderer and parser for CSS and JavaScript. The pages are looking very nice and it is easy to use (one class implementation to show a html page). I want to show HTML pages in swing which can show javascript generated and modified objects, like a flashing circle. As far as i know, the cobra toolkit can show static javascript commands like 'document.write' or closwWindow() implementations, but not updating the GUI from a JavaScript. Did someone works with Lobo Cobra Toolkit before and give me some sample code or tips for showing HTML pages with animated JavaScript objects inside.

    Read the article

  • Are Rich Snippets page specific or domain specific?

    - by John Isaacks
    Google will now parse certain microdata (for example reviews) on your web pages and display the info in search results. They call this Rich Snippets I am wondering is this page specific or domain specific? I keep all my reviews on a separate review page thats linked to from the home page. But my review page itself is very unlikely to be displayed in a search result, more likely to be displayed is my homepage or product landing page. But being that the review microdata is not on these pages (but is on the website). I am wondering if the rich snippets will be shown for these pages?

    Read the article

  • Page Footer not showing in Crystal Report

    - by Mike C.
    Hello, I am using Crystal Reports in Visual Studio 2008. I have about 5 pages worth of static text that needs to appear at the top of my report, so I put it in the report header section. I have a page footer section on the page that shows the page number. This does not show, and I suspect it has something to do with the long report header. How can I make the page footer show with a large report header? Edit: The Page Footer is actually appearing once on the last page. The Report Header takes up 5 pages and there isn't a page footer on any of those pages.

    Read the article

  • How to set permissions on SharePoint to hide an aspx page for authenticated users and to make it vis

    - by Flo
    I have a portal based on a publishing portal. The portal (SPSite) contains has two websites (SPWebs) one is anonymously accessible and other one isn't. This works as expected. Now I want to set the permissions for some aspx page of the anonymously accessible website so that they are not visible for authenticated users. So it's actually the opposite of anonymous access. User that are not logged in should see the aspx pages and logged in user shouldn't. The aspx pages are normal publishing pages of the publishing portal. How could I archive this. Is this possible at all?

    Read the article

< Previous Page | 69 70 71 72 73 74 75 76 77 78 79 80  | Next Page >