Search Results

Search found 12988 results on 520 pages for 'performance'.

Page 175/520 | < Previous Page | 171 172 173 174 175 176 177 178 179 180 181 182  | Next Page >

  • Can I expect a performance gain from removing this JOIN?

    - by makeee
    I have a "items" table with 1 million rows and a "users" table with 20,000 rows. When I select from the "items" table I do a join on the "users" table (items.user_id = user.id), so that I can grab the "username" from the users table. I'm considering adding a username column to the items table and removing the join. Can I expect a decent performance increase from this? It's already quite fast, but it would be nice to decrease my load (which is pretty high). The downside is that if the user changes their username, items will still reflect their old username, but this is okay with me if I can expect a decent performance increase. I'm asking stackoverflow because benchmarks aren't telling me too much. Both queries finish very quickly. Regardless, I'm wondering if removing the join would lighten load on the database to any significant degree.

    Read the article

  • How to improve performance of opening Microsoft Word when automated from c#?

    - by Abdullah BaMusa
    I have Microsoft Word template that I automated filling it’s fields from my application, and when the user request print I open this template. but creating word application every time user request print after filling fields is very expensive and lead to some delay while opening the template, so I choose to cache the reference to Word then just open the new filled template. that solve the performance issue as opening file is less expensive than recreating Word each time, but this work while the user just close the document not the entire Word application which when happened my reference to Word become invalid and return with exception says: “The RPC server is unavailable” next time request opening template . I tried to subscribe to BeforClosing event but his trigger for Quitting Word as well as Closing documents. My question is how to know if the word is closing document or quit the entire application so I take the proper action, or any hint for another direction of thinking about improve performance of opening word template.

    Read the article

  • C# performance of static string[] contains() (slooooow) vs. == operator

    - by Andrew White
    Hiya, Just a quick query: I had a piece of code which compared a string against a long list of values, e.g. if(str == "string1" || str = "string2" || str == "string3" || str = "string4". DoSomething(); And the interest of code clarity and maintainability I changed it to public static string[] strValues = { "String1", "String2", "String3", "String4"}; ... if(strValues.Contains(str) DoSomething(); Only to find the code execution time went from 2.5secs to 6.8secs (executed ca. 200,000 times). I certainly understand a slight performance trade off, but 300%? Anyway I could define the static strings differently to enhance performance? Cheers.

    Read the article

  • What are the differences in performance between synchronous and asynchronous JavaScript script loading?

    - by jasdeepkhalsa
    My question is simply: what are the differences in performance between synchronous and asynchronous JavaScript script loading? From what I've gathered synchronous code blocks the loading of a page and/or rest of the code from executing. This happens at two levels. First, at the level of the script actually loading, and secondly, within the JavaScript code itself. For example, on the page: Synchronous: <script src="demo_async.js" type="text/javascript"></script> Asynchronous: <script async src="demo_async.js" type="text/javascript"></script> And within a script: Synchronous: function a() {alert("a"); function b() {alert("b");}} Asynchronous (and self-executing): (function(a, function(b){ alert(b); }) { alert(a); }))(); So what really is the difference in performance from using these different loading methods and JavaScript patterns?

    Read the article

  • BI and EPM Landscape

    - by frank.buytendijk
    Most of my blog entries are not about Oracle products, and most of the latest entries are about topics such as IT strategy and enterprise architecture. However, given my background at Gartner, and at Hyperion, I still keep a close eye on what's happening in BI and EPM. One important reason is that I believe there is significant competitive value for organizations getting BI and EPM right. Davenport and Harris wrote a great book called "Competing on Analytics", in which they explain this in a very engaging and convincing way. At Oracle we have defined the concept of "management excellence" that outlines what organizations have to do to keep or create a competitive edge. It's not only in the business processes, but also in the management processes. Recently, Gartner published its 2009 market shares report for BI, Analytics, and Performance Management. Gartner identifies the same three segments that Oracle does: (1) CPM Suites (Oracle refers not to Corporate Performance Management, but Enterprise Performance Management), (2) BI Platform, and (3) Analytic Applications & Performance Management. According to Gartner, Oracle's share is increasing with revenue growing by more than 5%. Oracle currently holds the #2 market share position in the overall BI Software space based on total BI software revenue. Source: Gartner Dataquest Market Share: Business Intelligence, Analytics and Performance Management Software, Worldwide, 2009; Dan Sommer and Bhavish Sood; Apr 2010 Gartner has ranked Oracle as #1 in the CPM Suites worldwide sub-segment based on total BI software revenue, and Oracle is gaining share with revenue growing by more than 6% in 2009. Source: Gartner Dataquest Market Share: Business Intelligence, Analytics and Performance Management Software, Worldwide, 2009; Dan Sommer and Bhavish Sood; Apr 2010 The Analytic Applications & Performance Management subsegment is more fragmented. It has for instance a very large "Other Vendors" category. The largest player traditionally is SAS. Analytic Applications are often meant for very specific analytic needs in very specific industry sectors. According to Gartner, from the large vendors, again Oracle is the one who is gaining the most share - with total BI software revenue growth close to 15% in 2009. Source: Gartner Dataquest Market Share: Business Intelligence, Analytics and Performance Management Software, Worldwide, 2009; Dan Sommer and Bhavish Sood; Apr 2010 I believe this shows Oracle's integration strategy is working. In fact, integration actually is the innovation. BI and EPM have been silo technology platforms and application suites way too long. Management and measuring performance should be very closely linked to strategy execution, which is the domain of other business application areas such as CRM, ERP, and Supply Chain. BI and EPM are not about "making better decisions" anymore, but are part of a tangible action framework. Furthermore, organizations are getting more serious about ecosystem thinking. They do not evaluate single tools anymore for different application areas, but buy into a complete ecosystem of hardware, software and services. The best ecosystem is the one that offers the most options, in environments where the uncertainty is high and investments are hard to reverse. The key to successfully managing such an environment is middleware, and BI and EPM become increasingly middleware intensive. In fact, given the horizontal nature of BI and EPM, sitting on top of all business functions and applications, you could call them "upperware". Many are active in the BI and EPM space. Big players can offer a lot, but there are always many areas that are covered by specialty vendors. Oracle openly embraces those technologies within the ecosystem as well. Complete, open and integrated still accurately describes the Oracle product strategy. frank

    Read the article

  • How the number of indexes built on a table can impact performances?

    - by Davide Mauri
    We all know that putting too many indexes (I’m talking of non-clustered index only, of course) on table may produce performance problems due to the overhead that each index bring to all insert/update/delete operations on that table. But how much? I mean, we all agree – I think – that, generally speaking, having many indexes on a table is “bad”. But how bad it can be? How much the performance will degrade? And on a concurrent system how much this situation can also hurts SELECT performances? If SQL Server take more time to update a row on a table due to the amount of indexes it also has to update, this also means that locks will be held for more time, slowing down the perceived performance of all queries involved. I was quite curious to measure this, also because when teaching it’s by far more impressive and effective to show to attended a chart with the measured impact, so that they can really “feel” what it means! To do the tests, I’ve create a script that creates a table (that has a clustered index on the primary key which is an identity column) , loads 1000 rows into the table (inserting 1000 row using only one insert, instead of issuing 1000 insert of one row, in order to minimize the overhead needed to handle the transaction, that would have otherwise ), and measures the time taken to do it. The process is then repeated 16 times, each time adding a new index on the table, using columns from table in a round-robin fashion. Test are done against different row sizes, so that it’s possible to check if performance changes depending on row size. The result are interesting, although expected. This is the chart showing how much time it takes to insert 1000 on a table that has from 0 to 16 non-clustered indexes. Each test has been run 20 times in order to have an average value. The value has been cleaned from outliers value due to unpredictable performance fluctuations due to machine activity. The test shows that in a  table with a row size of 80 bytes, 1000 rows can be inserted in 9,05 msec if no indexes are present on the table, and the value grows up to 88 (!!!) msec when you have 16 indexes on it This means a impact on performance of 975%. That’s *huge*! Now, what happens if we have a bigger row size? Say that we have a table with a row size of 1520 byte. Here’s the data, from 0 to 16 indexes on that table: In this case we need near 22 msec to insert 1000 in a table with no indexes, but we need more that 500msec if the table has 16 active indexes! Now we’re talking of a 2410% impact on performance! Now we can have a tangible idea of what’s the impact of having (too?) many indexes on a table and also how the size of a row also impact performances. That’s why the golden rule of OLTP databases “few indexes, but good” is so true! (And in fact last week I saw a database with tables with 1700bytes row size and 23 (!!!) indexes on them!) This also means that a too heavy denormalization is really not a good idea (we’re always talking about OLTP systems, keep it in mind), since the performance get worse with the increase of the row size. So, be careful out there, and keep in mind the “equilibrium” is the key world of a database professional: equilibrium between read and write performance, between normalization and denormalization, between to few and too may indexes. PS Tests are done on a VMWare Workstation 7 VM with 2 CPU and 4 GB of Memory. Host machine is a Dell Precsioni M6500 with i7 Extreme X920 Quad-Core HT 2.0Ghz and 16Gb of RAM. Database is stored on a SSD Intel X-25E Drive, Simple Recovery Model, running on SQL Server 2008 R2. If you also want to to tests on your own, you can download the test script here: Open TestIndexPerformance.sql

    Read the article

  • Why don't SCOM R2 web console performance views load?

    - by Nexus
    When I select any performance view via my SCOM R2 web console, I get the following error: Unexpected error There was an error displaying the page you requested. ... and some suggestions about restarting my browser, which doesn't resolve the issue. The request produces the following event in the logs: Event code: 3005 Event message: An unhandled exception has occurred. Event time: 7/05/2010 11:41:38 AM Event time (UTC): 7/05/2010 1:41:38 AM Event ID: f4c47d1302694e1c8039e6c0088c2520 Event sequence: 18 Event occurrence:1 Event detail code: 0 [snip] Exception information: Exception type: HttpException Exception message: Error executing child request for /ResultViews/ViewTypePerformance.aspx. I'm using forms authentication and all other web console functionality works perfectly. My server is Windows 2008 R2 Standard running SCOM R2 and runs the DB, Web Console and RMS roles. Has anyone else experienced this issue? Is it fixed in the cumulative update release for SCOM R2?

    Read the article

  • What's the performance on USB docking stations, and can they be used when laptop is closed?

    - by David
    I'm looking into a docking station for a Dell Studio laptop. I don't see the traditional docking stations I'm familiar with - the kind (for a Dell Latitude, for example) where you sit the laptop on top of a long row of pins. Instead, I'm seeing a lot of USB docking stations. When I close my laptop, I want it to go into sleep mode. If I then connect a USB docking station to the laptop while it's closed, will it wake up? What's the performance on USB 2.0 docking stations with a new Dell Studio? Can all of the video and internet traffic really go through a USB 2.0 connection while still providing the best video frame rates and internet speeds? When you undock, I assume you'd have to use the "Safely Remove Hardware" feature in Windows. Will that successfully 'remove' everything attached to the docking station - external drives, thumb drives, etc?

    Read the article

  • Server performance worsened after a hardware upgrade: how should I reconfigure the server?

    - by twick
    I'm running a site on an Ubuntu/Apache/Django/PostgreSQL stack. We upgraded our server recently from 1 processor with 2 Gb total RAM (with 0.5 Gb of that RAM assigned to memcached) to a new server that has 2 processors with 4 Gb total RAM (with 2 Gb of that RAM assigned to memcached). However, when I looked at Google Webmaster Tools, I found out that the average page speed has worsened from 5 seconds to 15 seconds. Why would performance get worse with a hardware upgrade? What should I check and tune? Is this more likely to be a problem with memcached, Apache, Django, or PostgreSQL?

    Read the article

  • Maximum number of files in one ext3 directory while still getting acceptable performance?

    - by knorv
    I have an application writing to an ext3 directory which over time has grown to roughly three million files. Needless to say, reading the file listing of this directory is unbearably slow. I don't blame ext3. The proper solution would have been to let the application code write to sub-directories such as ./a/b/c/abc.ext rather than using only ./abc.ext. I'm changing to such a sub-directory structure and my question is simply: roughly how many files should I expect to store in one ext3 directory while still getting acceptable performance? What's your experience? Or in other words; assuming that I need to store three million files in the structure, how many levels deep should the ./a/b/c/abc.ext structure be? Obviously this is a question that cannot be answered exactly, but I'm looking for a ball park estimate.

    Read the article

  • Advise about performance for local or remote SQL Server?

    - by TruMan1
    I currently have my web server and SQL Express / MySQL server on the same server. It is on a VPS. I have been having problems with my hosting so I am thinking of separating the web and db server into 2 VPS servers. Does anyone recommend this? I am worried that changing my setup from a local DB server to a remote one will degrade performance heavily. They will not be on the same network, but will reference each other via an IP address. Anything I should be aware of?

    Read the article

  • Move virtual machine hard disk to a separate physical hard disk for better performance?

    - by joeeoj
    I have a dual-core machine with the host OS and many guest virtual OSs. Although I have 8GB of RAM, I notice a slowdown when I turn some virtual machine on (and it takes only 1GB RAM). I was told that I should move virtual machine hard disk file to a separate (another) physical hard drive in my PC to get better performance. This way the head of the hard disk would not have to jump from the virtual OS to the host OS as each hard drive would have its own head to deal with the OS: hard drive 1 head for host OS and hard drive 2 head for guest OS. Is this true? Should I get another hard disk only for virtual machine hard disk files?

    Read the article

  • Does disable log error for MySQL increasing it's performance ? How disable it?

    - by adnan
    Does disable log error for MySQL increasing it's performance ? How disable it ? This is my service status Server load 0.63 (8 CPUs) Memory Used 23.38% (957,600 of 4,096,000) Swap Used 0% (0 of 1) And this is print screen for process manager http://elnhrda.com/promgr.jpg This is my.cnf [mysqld] query_cache_size=64M skip-name-resolve #innodb_file_per_table=1 query_cache_limit=2M read_buffer_size = 2M read_rnd_buffer_size = 16M sort_buffer_size = 8M join_buffer_size = 8M thread_cache_size = 8 thread_concurrency = 8 innodb_buffer_pool_size = 2G Iam looking for doing any thing to increase my website speed I have VPS 4G.B RAM CENTOS 6 X86_64 Note please : this statics taken now which no any queries executed & site have not any visitors in the same time

    Read the article

  • For Intel cpu , if chipset and motherboard are also from intel then it will give best performance. I

    - by metal gear solid
    I'm going to purchase Intel® Core™2 Duo Processor E7500 (3M Cache, 2.93 GHz, 1066 MHz FSB) and for motherboard my local vendor suggesting me to purchase Intel DG41RQ MB motherboard and he is also saying if i'm purchasing Intel CPU then purchasing Intel's Own motherboard with intel chipset will give best performance. Is it true? To get good inbuilt graphic I'm thinking to purchase nvidia chipset based motherboard of any other company like Asus, Gigabyte, MSi etc. is it ok? Although i never play games on my PC but thinking Inbuilt Nvidia graphics will be better for running Photoshop and watching movies then Intel's inbuilt graphics. or it's ok to purchase Intel DG41RQ MB motherborad as suggested by local vendor. Intel's inbuilt graphics would be enough for Photoshop and Watching movies. If you know any other good motherboard for Intel® Core™2 Duo Processor E7500 (3M Cache, 2.93 GHz, 1066 MHz FSB) then tell.

    Read the article

  • Will disabling NAT on my wireless router cause a performance hit on my DSL router?

    - by user117313
    I have a Thomson TG508v2 router/modem and a brand new AirPort Extreme connected to it. At first, I set the AirPort Extreme to "share a public IP", which enables DHCP/NAT. Everything works great but I was having the Double NAT error. So I set the AirPort Extreme as bridge mode, which disables DHCP/NAT and let my crappy modem handle my internal network IPs. Will this cause any performance hit in my network? I'm worried because I'll outsource this job from AirPort Extreme (premium hardware) to the crappy modem. Before you suggest, I tried setting my modem as a bridge and configuring PPPoE on the AirPort Extreme, however it wouldn't connect to the internet, so I just dismissed leaving the modem as bridge.

    Read the article

  • Is it possible to combine two internet connections to increase performance?

    - by cornjuliox
    I've got a small home network, 3 PCs plus a laptop or two when the relatives come to visit, connected to a single cable internet connection. Now, as soon as everyone starts using the 'net the performance starts to suffer and if the load is heavy enough nobody can get anything done and everyone complains. At one point it was so bad that only one of us could use it at a time. I was researching possible solutions to this problem and I heard that internet cafes that utilize 2 internet connections, possibly from different providers, and have some sort of router that allows them to split the traffic between the both of them, with online games going through one and web traffic going through another. Is this possible? What is the technical term for it, and can/should it be applied to a home network setup or is there another solution to this problem?

    Read the article

  • Can I use CAT 6a connectors with 7a cable (and get 6a performance)?

    - by Mr. Flibble
    I'm re-wiring a building and want to get the best cable possible laid - it required re-plastering to make a change to the cables and the cables will be there for the next 10 - 20 years. Currently there appears to be cat 7a cable available but not too much in the way of cat 7a connectors. Also - I won't be using 40Gig hardware in the near future. So, my question: is it possible to use cat 6a connectors / patch panels with cat 7a cable and get the same performance as I would had I used cat 6a cable? Are there any gotchas in trying to do this?

    Read the article

  • Doesn't DNS diversity negatively affect performance? Why/how?

    - by cnst
    If you look at the press releases of various orgs that run the internet, you can see them praise the fact that now they run root server X in city Y, as if that magically makes everyone in city Y get all the relevant resolutions from the local server X, instead of going 200ms across the oceans and lands to other continents for resolutions. Similarly, the zones of some geographical domain names, like .ru, are being mirrored not just within Europe, but also, for example, in Hong Kong, which is no more, no less, but is about 300ms away from central Europe, since the traffic is often crossing the two oceans on each way. Doesn't all of this negatively affect DNS performance? Isn't it more of a liability to have a diverse pool of geodispersed authoritative servers, especially if your target audience is quite geographically concentrated? Perhaps a better question is, are there any DNS resolvers that use something better than the naive round-robin for choosing which authoritative server to contact?

    Read the article

  • Does using VLANs in your network infrastructure cause an appreciable decrease in performance?

    - by Peter Grace
    This is something I've never considered before and wanted the opinions of the experts. We use VLANs day in and day out for various network tasks. My modus operandi is that in general, if something supports VLANs, that port is getting trunked because it just makes a ton of sense if there's even the slightest chance you need to do more than one thing on that single link. As I ponder this, though, I'm wondering whether there's a performance penalty involved with this line of thinking? Is the impact negligible?

    Read the article

  • Is an I/O benchmark made for hardware an accurate assessment of a Windows VM's performance under vSphere 5?

    - by Jeremy
    We support an enterprise application running on Windows Server 2008 R2. One of our customers has chosen to install to VMWare, and what I'm finding is that the VM's are relatively slow compared to hardware. Our product development team has advised that many VMs appear to run particularly slow on I/O benchmarks, which impact performance in production. I've tried the AttoSoft I/O benchmark and find that for smaller I/O blocks (1-32K) the VM I'm looking at is 25x slower than hardware and for larger I/O blocks (1-8MB) it's 10x slower. Is this a fair benchmark? If not, any suggestions for a fair test?

    Read the article

  • Will installing an Ultra ATA cable backwards affect performance?

    - by GMMan
    I've recently purchased a hard drive upgrade for my Xbox 320GB WD Caviar Blue WD3200AAJB and StarTech.com Ultra ATA/66/100/133 cable IDE66 yes I'm crazy When it came to installing the cable, it was too short (my fault), and there wasn't enough space between the master and slave ends to reach both the DVD drive and the hard drive. The only thing I could do was install the cable backwards and twisting it quite a bit to make it fit. The upgrade works, but reading the manual for the hard drive I replaced (10GB Seagate U Series 5), apparently there is a specific way you have to connect the cable. I don't have that option, so the question comes down to, will my drive performance be at Ultra ATA levels, or is it still performing at original ATA speeds? Is there any way I can test this (benchmarking software for Xbox)?

    Read the article

  • Windows-Vista events: Diagnostics-Performance: How-To read this information?

    - by Ice
    Hi, i am wondering how long the bootprocess needs and looking in %SystemRoot%\system32\eventvwr.msc /s. Some entries marked as critical like : Starting needs : 184707ms, sometimes 211855ms or 269767ms Some Errors like: This process does many diskactivities and lower the performance of Windows: Filename : ntoskrnl.exe Why are the startups marked as critical? Are the values normal on a Dell Precision M90 (Intel Centrono Dual CPU, 2GB RAM, 80 GB Disk)? Some entries marked as Errors showing the time for shutdown, but there are some like this one i printed in this question, what is the meaning of this?

    Read the article

  • Performance problems - Jira running on Ubuntu over VMware ESX 4.0 maxing out all 4 vCPUs.

    - by Jack T
    We are running Jira on a box under VMware ESX 4.0 and performance is vaiable to say the least. The physical box has 12 Gig RAM and 4x Xeon 2.26 GHz CPUs. vCentre is telling us the CPUs are not maxed out at any time, RAM is fine too. When we issue a request to the host it sometimes maxes out all 4 vCPUs. Sometimes it's quick, sometimes very very slow. There doesn't seem to be a pattern. Any ideas?

    Read the article

< Previous Page | 171 172 173 174 175 176 177 178 179 180 181 182  | Next Page >