Search Results

Search found 30217 results on 1209 pages for 'website performance'.

Page 100/1209 | < Previous Page | 96 97 98 99 100 101 102 103 104 105 106 107  | Next Page >

  • Slow Internet Performance in 12.04 LTS

    - by Mad
    Have installed Ubuntu 12.04 LTS. I have encountered the below problems. 1. Have two OS. Internet is too slow in U 12.04 compared to Windows 7 2. System Performance is very slow 3. After installing Ubuntu 12.04, my brightness is dark during the initiail time. However, I have resolved this issue and found to be working fine. 4. Unable to connect Wireless network after inputing security credentials. Please note, I am beginner to this Linux. Would Appreciate if someone could explain in step by step to overcome the above issues.

    Read the article

  • Developing Schema Compare for Oracle (Part 6): 9i Query Performance

    - by Simon Cooper
    All throughout the EAP and beta versions of Schema Compare for Oracle, our main request was support for Oracle 9i. After releasing version 1.0 with support for 10g and 11g, our next step was then to get version 1.1 of SCfO out with support for 9i. However, there were some significant problems that we had to overcome first. This post will concentrate on query execution time. When we first tested SCfO on a 9i server, after accounting for various changes to the data dictionary, we found that database registration was taking a long time. And I mean a looooooong time. The same database that on 10g or 11g would take a couple of minutes to register would be taking upwards of 30 mins on 9i. Obviously, this is not ideal, so a poke around the query execution plans was required. As an example, let's take the table population query - the one that reads ALL_TABLES and joins it with a few other dictionary views to get us back our list of tables. On 10g, this query takes 5.6 seconds. On 9i, it takes 89.47 seconds. The difference in execution plan is even more dramatic - here's the (edited) execution plan on 10g: -------------------------------------------------------------------------------| Id | Operation | Name | Bytes | Cost |-------------------------------------------------------------------------------| 0 | SELECT STATEMENT | | 108K| 939 || 1 | SORT ORDER BY | | 108K| 939 || 2 | NESTED LOOPS OUTER | | 108K| 938 ||* 3 | HASH JOIN RIGHT OUTER | | 103K| 762 || 4 | VIEW | ALL_EXTERNAL_LOCATIONS | 2058 | 3 ||* 20 | HASH JOIN RIGHT OUTER | | 73472 | 759 || 21 | VIEW | ALL_EXTERNAL_TABLES | 2097 | 3 ||* 34 | HASH JOIN RIGHT OUTER | | 39920 | 755 || 35 | VIEW | ALL_MVIEWS | 51 | 7 || 58 | NESTED LOOPS OUTER | | 39104 | 748 || 59 | VIEW | ALL_TABLES | 6704 | 668 || 89 | VIEW PUSHED PREDICATE | ALL_TAB_COMMENTS | 2025 | 5 || 106 | VIEW | ALL_PART_TABLES | 277 | 11 |------------------------------------------------------------------------------- And the same query on 9i: -------------------------------------------------------------------------------| Id | Operation | Name | Bytes | Cost |-------------------------------------------------------------------------------| 0 | SELECT STATEMENT | | 16P| 55G|| 1 | SORT ORDER BY | | 16P| 55G|| 2 | NESTED LOOPS OUTER | | 16P| 862M|| 3 | NESTED LOOPS OUTER | | 5251G| 992K|| 4 | NESTED LOOPS OUTER | | 4243M| 2578 || 5 | NESTED LOOPS OUTER | | 2669K| 1440 ||* 6 | HASH JOIN OUTER | | 398K| 302 || 7 | VIEW | ALL_TABLES | 342K| 276 || 29 | VIEW | ALL_MVIEWS | 51 | 20 ||* 50 | VIEW PUSHED PREDICATE | ALL_TAB_COMMENTS | 2043 | ||* 66 | VIEW PUSHED PREDICATE | ALL_EXTERNAL_TABLES | 1777K| ||* 80 | VIEW PUSHED PREDICATE | ALL_EXTERNAL_LOCATIONS | 1744K| ||* 96 | VIEW | ALL_PART_TABLES | 852K| |------------------------------------------------------------------------------- Have a look at the cost column. 10g's overall query cost is 939, and 9i is 55,000,000,000 (or more precisely, 55,496,472,769). It's also having to process far more data. What on earth could be causing this huge difference in query cost? After trawling through the '10g New Features' documentation, we found item 1.9.2.21. Before 10g, Oracle advised that you do not collect statistics on data dictionary objects. From 10g, it advised that you do collect statistics on the data dictionary; for our queries, Oracle therefore knows what sort of data is in the dictionary tables, and so can generate an efficient execution plan. On 9i, no statistics are present on the system tables, so Oracle has to use the Rule Based Optimizer, which turns most LEFT JOINs into nested loops. If we force 9i to use hash joins, like 10g, we get a much better plan: -------------------------------------------------------------------------------| Id | Operation | Name | Bytes | Cost |-------------------------------------------------------------------------------| 0 | SELECT STATEMENT | | 7587K| 3704 || 1 | SORT ORDER BY | | 7587K| 3704 ||* 2 | HASH JOIN OUTER | | 7587K| 822 ||* 3 | HASH JOIN OUTER | | 5262K| 616 ||* 4 | HASH JOIN OUTER | | 2980K| 465 ||* 5 | HASH JOIN OUTER | | 710K| 432 ||* 6 | HASH JOIN OUTER | | 398K| 302 || 7 | VIEW | ALL_TABLES | 342K| 276 || 29 | VIEW | ALL_MVIEWS | 51 | 20 || 50 | VIEW | ALL_PART_TABLES | 852K| 104 || 78 | VIEW | ALL_TAB_COMMENTS | 2043 | 14 || 93 | VIEW | ALL_EXTERNAL_LOCATIONS | 1744K| 31 || 106 | VIEW | ALL_EXTERNAL_TABLES | 1777K| 28 |------------------------------------------------------------------------------- That's much more like it. This drops the execution time down to 24 seconds. Not as good as 10g, but still an improvement. There are still several problems with this, however. 10g introduced a new join method - a right outer hash join (used in the first execution plan). The 9i query optimizer doesn't have this option available, so forcing a hash join means it has to hash the ALL_TABLES table, and furthermore re-hash it for every hash join in the execution plan; this could be thousands and thousands of rows. And although forcing hash joins somewhat alleviates this problem on our test systems, there's no guarantee that this will improve the execution time on customers' systems; it may even increase the time it takes (say, if all their tables are partitioned, or they've got a lot of materialized views). Ideally, we would want a solution that provides a speedup whatever the input. To try and get some ideas, we asked some oracle performance specialists to see if they had any ideas or tips. Their recommendation was to add a hidden hook into the product that allowed users to specify their own query hints, or even rewrite the queries entirely. However, we would prefer not to take that approach; as well as a lot of new infrastructure & a rewrite of the population code, it would have meant that any users of 9i would have to spend some time optimizing it to get it working on their system before they could use the product. Another approach was needed. All our population queries have a very specific pattern - a base table provides most of the information we need (ALL_TABLES for tables, or ALL_TAB_COLS for columns) and we do a left join to extra subsidiary tables that fill in gaps (for instance, ALL_PART_TABLES for partition information). All the left joins use the same set of columns to join on (typically the object owner & name), so we could re-use the hash information for each join, rather than re-hashing the same columns for every join. To allow us to do this, along with various other performance improvements that could be done for the specific query pattern we were using, we read all the tables individually and do a hash join on the client. Fortunately, this 'pure' algorithmic problem is the kind that can be very well optimized for expected real-world situations; as well as storing row data we're not using in the hash key on disk, we use very specific memory-efficient data structures to store all the information we need. This allows us to achieve a database population time that is as fast as on 10g, and even (in some situations) slightly faster, and a memory overhead of roughly 150 bytes per row of data in the result set (for schemas with 10,000 tables in that means an extra 1.4MB memory being used during population). Next: fun with the 9i dictionary views.

    Read the article

  • First Project a big one, How much should we charge?

    - by confuzzled
    Two of my cousins and I started a freelance computer repair/web design business just to make some money on the side during college, and received our first major web design project about three weeks ago. Now we've created websites before, but it was mostly for family businesses and have never really charged money, and most of the websites have been static, and don't really require a CMS. This project, however, was a big one (for us anyways). We created a news site that had several categories, we created the banners, we created a classifieds page (not a web app just something static that they control). Several links, a few graphical assets, CSS drop down menu, RSS feed from a different news site, weather, all the normal stuff you would find on a regular news site. On top of that we put in all the usual Joomla stuff (search, Jcomments, Jslide pictures, JCE, etc.). Then we uploaded the first 10 articles they gave us, and we are going to train them how to use Joomla. Now, at first we decided for 700 dollars. I assumed they just wanted a simple blog like website where they can upload articles. But then we had a meeting, and they asked for a lot more. Note: we did not hard code the template from scratch, but customized the gantry framework to fit their needs. We did code quite a bit however. I estimate that we put in about 50-60 hours in total. I'm wondering if 700 dollars is a bit low, this price is definitely not set in stone. Please keep in mind that this is our first project, and we are newbies, please be kind. Thank You!

    Read the article

  • Radeon HD 6290 terrible performance on a certified laptop

    - by dac
    I bought Asus K53U laptop, which is Ubuntu certified with pre-installed 11.10. The graphic card is Radeon HD 6290 but 720p playback is terrible. Even page scrolling in Firefox is very laggy. Proprietary drivers are installed by default. How is this possible, why is the laptop Ubuntu certified if the performance is poor? Any solution to this? I just did apt-get autoremove, and after that, this message came out in terminal: Error inserting vesafb (/lib/modules/3.0.0-15-generic/kernel/drivers/video/vesafb.ko): No such device Could that be the problem?

    Read the article

  • Android performance/issues with Corona SDK?

    - by B5Fan74
    I know this is a fairly broad question. We are looking to develop a mobile game and want to use a multi-platform engine/SDK. We like what we see with Corona but in doing some reading, we are seeing a lot of references to poor performance on the 'droid platforms. I am unsure how much of this is still relevant? Many of the articles/posts/references/discussions vary in date from 18 months ago to earlier this year. Is there a reason we should not pursue Corona if Android support is important to us? The game is going to be 2D isometric view. Thanks!

    Read the article

  • Logging in JSON Effect on Performance

    - by Pius
    I see more and more articles about logging in JSON. You can also find one on NodeJS blog. Why does everyone like it so much? I can only see more operations getting involved: A couple new objects being created. Stringifying objects, which either involves calculating string length or multiple string allocations. GCing all the crap that was created. Is there any test on performance when using JSON logging and regular string logging? Do people use JSON (for logging) in enterprise projects?

    Read the article

  • ASP.NET Menu, NavBar And Pager Performance Improvements v2010 vol 1

    Check out the improvements weve made to some of our ASP.NET controls in the DXperience v2010.1 release. We changed the rendering of our ASP.NET AJAX Menu, Navigation Pane and Pager controls. The controls now use semantic rendering combined with advanced CSS styles, which results in a dramatic decrease of HTML output, improved performance and a reduction in the servers workload. Also, several of our other ASP.NET controls like the ASPxGridView and ASPxScheduler also benefit because The primary...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • SQLAuthority News Microsoft SQL Server 2005/2008 Query Optimization & Performance Tuning Training

    Last 3 days to register for the courses. This is one time offer with big discount. The deadline for the course registration is 5th May, 2010. There are two different courses are offered by Solid Quality Mentors 1) Microsoft SQL Server 2005/2008 Query Optimization & Performance Tuning – Pinal Dave Date: May 12-14, 2010 Price: [...]...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Google I/O 2012 - High Performance HTML5

    Google I/O 2012 - High Performance HTML5 For years we built web apps that far outpaced the capabilities of the browsers they ran in. Just as the browsers were catching up HTML5 came on the scene - video and audio, canvas, SVG, app cache, localStorage, @font-face, and more. Now the browsers are racing to stay ahead of the wave that's building as developers adopt these new capabilities. Is your HTML5 app going to ride the wave or be dashed on the rocks leaving users stranded? Learn which HTML5 features to seek out and which to avoid when it comes to building fast HTML5 web apps. This session will be live captioned. From: GoogleDevelopers Views: 91 1 ratings Time: 01:02:07 More in Science & Technology

    Read the article

  • ASP.NET Performance, 100 "Memory Hard Faults" indiciate a memory swapping problem?

    - by Robert
    With a customer web site we currently experiences performance problems. While analyzing the problem we found an unexpected amount of of 112 "Memory Hard Faults" per minute. Does anybody can interpret the meaning of this value? Does this happen, when memory swapping is necessary - so the root cause is not sufficient memory? Even if the CPU value seems high, it is not the main problem for the slow web site. Do you agree?

    Read the article

  • problems with graphic performance in Ubuntu 12.04

    - by Falk
    in advance: my English is not perfect ;) processor: Intel® Core™2 Duo CPU E8500 @ 3.16GHz × 2 memory: 3,9 GiB graphics: GeForce GTX 460 Ubuntu 12.04 32-Bit fresh installed (previously 11.10 32-Bit) driver: NVIDIAs accelerated Driver (Version current-updates) probelm: In Ubuntu 12.04 I have problems with the performance of the graphics (games for example: Volley Brwawl, Neverball, Beep and Minecraft), and effects of windows (always at minimize, maximize...) There are laggs at animations, and the laggs in games are extremely annoying. Problems occur only in Ubuntu 12.04, everything was good in 11.10. No Problems with YouTube videos, HD videos. Is there a solution? BBecause I have found nothing here in the Forum and on Google. Or is an update coming soonfor driver or whatever? (This Bug is already registered in Launchpad.) Thank you!

    Read the article

  • ASP.NET Combo Box and List Box Performance Improvements - v2010 vol 1

    Check out this great new performance feature of our ASP.NET combo box and list box controls for the DXperience v2010.1 release. You can now manually populate lists with items based on the currently applied filter criteria. This means that you can significantly decrease web server workload by loading only a subset of all items when working with large datasets. For instance, when using a large data source, you can only request a few records to be visible on the screen. The rest of the items can...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Premier Support for Hyperion Enterprise Performance Management System 11.1.1.x Ends July 2013

    - by inowodwo
    Premier Support for Oracle Hyperion Enterprise Performance Management System release 11.1.1.x expires July 2013. After July 2013, Sustaining Support will continue to be provided in accordance with Oracle's Lifetime Support Policy. Customers must follow a supported upgrade path.If your deployment is at EPM System release 11.1.1.4- Your supported upgrade path is directly to release 11.1.2.2.If your deployment is at EPM System Release 11.1.1.3 release- Your supported upgrade path is directly to release 11.1.2.2.If your deployment is at a prior 11.1.1.3 release- The recommended path is to upgrade to release 11.1.1.3. From here, you can continue a direct upgrade to release 11.1.2.2. For more information see Doc ID 1511588.1 or the Oracle Lifetime Support Policy

    Read the article

  • Performance Driven Manufacturing

    Manufacturers are searching for new, creative ways to address growing demands of global manufacturing. They want the latest tools and technologies to boost performance from their operations, suppliers, partners, distributors, and extended ecosystem, and they need global views for better visibility - both internally and across the extended supply chain. In addition, operations must move information more effectively to gain real-time insight into manufacturing shop floor status. Whether it's inside the plant or outside the traditional factory walls, manufacturers are searching for solutions to help them produce more for less, lower their total cost of ownership (TCO), and improve their return on investment (ROI).

    Read the article

  • How to handle new domain names?

    - by michael
    I have a new product which I'll call a pen ink reloader. I have a website using my products name, for example, www.inkywink.com which I want to have accessed by searches for keywords such as "pen ink", "pen out of ink" "ink for pens" etc. , since nobody knows that a pen ink reloader exists. I see that its quite difficult to get on front page for these keywords since they have lots of competition. However I notice that the exact phrases I want to rank highly for are available as domains. I purchase "www.penink.com" and "penoutofink.com" which for arguments sake are highly searched and the perfect keywords to get eyes on my money site www.inkywink.com . Two questions: 1. What is my best option to leverage those names so that they appear near top of searches so that I can get traffic to my money site? Do I just have them redirect 301 to inkywink.com or should I create small original content on each with links to my main site? 2. If I just have them redirected to inkywink.com, am I able to use keywords in metatag and headers for each site separately or do they all automatically obtain the same headers and tags as the site to which theyre redirected ? Thanks to anyone who can help as I'm a real newbie to all this.

    Read the article

  • ubuntu 14.04 - only performance and powersave governor available (i5 4.gen)

    - by kinske
    After updating my ubuntu 13.10 to 14.04 there only these to govenors available for my cpu, instead of additionally ondemand and conservative. cat /sys/devices/system/cpu/cpu0/cpufreq/scaling_available_governors performance powersave My cpu is a Intel(R) Core(TM) i5-4200M CPU @ 2.50GHz, my laptop a LENOVO B5400 80B6QB0 and my kernel version is 3.14.1-031401-generic x86_64 Before updating the four governors were available. If this is important, I have installed tlp and indicator-cpufreq. How do I get back the missing two govenors? ondemand is really important to me because of power saving.

    Read the article

  • Why Should I Avoid Inline Scripting?

    - by thesunneversets
    A knowledgeable friend recently looked at a website I helped launch, and commented something like "very cool site, shame about the inline scripting in the source code". I'm definitely in a position to remove the inline scripting where it occurs; I'm vaguely aware that it's "a bad thing". My question is: what are the real problems with inline scripting? Is there a significant performance issue, or is it mostly just a matter of good style? Can I justify immediate action on the inline scripting front to my superiors, when there are other things to work on that might have a more obvious impact on the site? If you pulled up to a website, and took a peek at the source code, what factors would lead you to say "hmm, professional work here", and what would cause you to recoil from an obviously amateurish job? Okay, that question turned into multiple questions in the writing. But basically, inline scripting - what's the deal?

    Read the article

  • SQL SERVER Out of the Box Activty and Performance Reports from SSSMS

    SQL Server management Studio 2008 is wonderful tool and has many different features. Many times, an average user does not use them as they are not aware about these features. Today, we will learn one such feature. SSMS comes with many inbuilt performance and activity reports, but we do not use it to the full [...]...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • poor performance while exporting data to excel from DB2

    - by Naga
    I am facing a performance issue while exporting data from DB2 to Excel 2003. Well the very first reason is file is about 10+ MB where it goes outofMemory Exception. I am using XLSTransformer and HSSFWorkbook classes to transform my xls file. I also have joins in my query( optional). But user most likely is going to choose these options. When they do so, of course, the data becomes huge and take lot of time and some times goes outOfMemory too. So Please advice me on this.

    Read the article

  • Performance impact of the new mtmalloc memory allocator

    - by nospam(at)example.com (Joerg Moellenkamp)
    I wrote at a number of occasions (here or here), that it could be really beneficial to use a different memory allocator for highly-threaded workloads, as the standard allocator is well ... the standard, however not very effective as soon as many threads comes into play. I didn't wrote about this as it was in my phase of silence but there was some change in the allocator area, Solaris 10 got a revamped mtmalloc allocator in version Solaris 10 08/11 (as described in "libmtmalloc Improvements"). The new memory allocator was introduced to Solaris development by the PSARC case 2010/212. But what's the effect of this new allocator and how does it works? Rickey C. Weisner wrote a nice article with "How Memory Allocation Affects Performance in Multithreaded Programs" explaining the inner mechanism of various allocators but he also publishes test results comparing Hoard, mtmalloc, umem, new mtmalloc and the libc malloc. Really interesting read and a must for people running applications on servers with a high number of threads.

    Read the article

  • Temenos WealthManager: performance benchmark on Exadata

    - by Javier Puerta
    Temenos WealthManager is at work in financial institutions of all sizes. No matter the size, each of Temenos’ customers has one requirement in common: a need for fast technology. This led Temenos to conduct a performance benchmark, running its WealthManager platform on Oracle Exadata Database Machine. During the study, Temenos executed high-intensity financial engines two- to three-times faster than its previous internal benchmarks and/or largest customer throughput. The company also demonstrated greater than two-times the average improvement in response times for six-times the number of users and data volumes. Further, Temenos secured a two-times gain in service-level agreements for batch and user-oriented workloads via dedicated or parallelized processing windows. Last year Temenos also communicated the availability of its T24 core banking system and how the benchmark run demonstrated the ability of Oracle’s Exadata platform to comfortably support the highest banking volumes for T24. Read full story here  

    Read the article

< Previous Page | 96 97 98 99 100 101 102 103 104 105 106 107  | Next Page >