Search Results

Search found 20904 results on 837 pages for 'disk performance'.

Page 486/837 | < Previous Page | 482 483 484 485 486 487 488 489 490 491 492 493  | Next Page >

  • I loose some directories when i upgrade from Ubuntu 11.10 to 12.04

    - by maythux
    last day i upgraded my ubunut 11.10 desktop to ubuntu 12.04. I was running a KVM virtual about 7 machines and managed by virt-manage software.... anyway when i finished upgrading i found that virt-manager is not working so i have to reconfigure it again and install some other missing packages that was deleted!!!! anyway i solve this issue...then i started to restore my virtual machines i restore 2 machines without any problems the third and fourth ones (windows) make a check disk that takes more that 6 hours but finally it works... other machines i cant find their attached hard disks i don't know what happens but i cant found that files. 1- upgrading delete files!?!! 2- Is there anyway to restore those files? thanks in advance

    Read the article

  • New customer references for Exadata-based projects

    - by Javier Puerta
    Milletech Systems, Inc. shows a large state university how to improve query response times 15-fold using its grant management solution built on Oracle’s Extreme Performance Infrastructure. Read More. Ação Informática helped Valdecard realize a 15-Fold Improvement in Fleet-Management and Benefit Card Data Processing using Oracle Exadata Machine. Read More. Neusoft deployed Benxi Municipal Human Resources and Social Security Bureau’s cloud-based database platform to process social insurance payments 50-times faster using Oracle Exadata Database Machine. Read More.

    Read the article

  • Windows Azure CDN(Content Delivery Network)

    - by kaleidoscope
    Windows Azure CDN caches your Windows Azure blobs at strategically placed locations to provide maximum bandwidth for delivering your content to users. You can enable CDN delivery for any storage account via the Windows Azure Developer Portal. The CDN provides edge delivery only to blobs that are in public blob containers, which are available for anonymous access. Windows Azure CDN has 18 locations globally (United States, Europe, Asia, Australia and South America) and continues to expand. The benefit of using a CDN is better performance and user experience for users who are farther from the source of the content stored in the Windows Azure Blob service. In addition, Windows Azure CDN provides worldwide high-bandwidth access to serve content for popular events. Current CDN locations in US. For more details please refer to the link.  http://blogs.msdn.com/windowsazure/archive/2009/11/05/introducing-the-windows-azure-content-delivery-network.aspx Sarang

    Read the article

  • Dashboard to aggregate Google Analytics, Facebook, YouTube etc tracking data?

    - by Richard
    I'd like to see as much tracking data as possible about my online presence, in one single dashboard - so views/conversions from Google Analytics data, the performance of my Facebook campaigns via the Insights API, views/clicks from my YouTube campaigns, etc. This could be as simple as a graph with time on the x-axis, and key indicators from each source on the y-axis (conversions from Analytics, likes on Facebook, views on YouTube, etc). The idea is that I can see customer engagement with each source, over time. I can write my own such dashboard easily enough, but I wondered if there was something off-the-shelf that already did this. Apologies if this isn't the right forum for such a question - would appreciate tips for the best place to ask.

    Read the article

  • Futures/Monads vs Events

    - by c69
    So, the question is quite simple: in an application framework, when performance impact can be ignored (10-20 events per second at max), what is more maintainable and flexible to use as a preferred medium for communication between modules - Events or Futures/Promices/Monads ? Its often being said, that Events (pub/sub, mediator) allow loose-coupling and thus - more maintainable app... My experience deny this: once you have more that 20+ events - debugging becomes hard, and so is refactoring - because it is very hard to see: who, when and why uses what. Promices (i'm coding in javascript) are much uglier and dumber, than Events. But: you can clearly see connections between function calls, so application logic becomes more straight-forward. What i'm afraid. though, is that Promices will bring more hard-coupling with them... p.s: the answer does not have to be based on JS, experience from other functional languages is much welcome.

    Read the article

  • Getting Started with Oracle Fusion Governance, Risk and Compliance (GRC)

    Designed from the ground-up using the latest technology advances and incorporating the best practices gathered from Oracle's thousands of customers, Fusion Applications are 100 percent open standards-based business applications that set a new standard for the way we innovate, work and adopt technology. Delivered as a complete suite of modular applications, Fusion Applications work with your existing portfolio to evolve your business to a new level of performance. In this AppCast, part of a special series on Fusion Applications, you hear about the unique advantages of Oracle Fusion Governance, Risk and Compliance and discover how Fusion GRC works with your existing applications investments.

    Read the article

  • Bit copy of encrypted home and other partitions

    - by Mka
    My laptop is overheating so I need to save all my files before I format the hard drive. I learned how to copy dev/sdX using dd command. However, I am not sure what to copy. Picture from GParted here: http://is.muni.cz/www/256590/fig.png should I copy sda5 and sda6 only? Or sda2 and sda1? I do not need to use these data on another disk, I just want to be able to access them - therefore I want to put them on external hard drive. And last question - how I will then mount my encrypted home? Will it work? Thanks a lot!

    Read the article

  • eSeminar ISV Partner Update: High Quality Reporting for Your Applications

    - by Mike.Hallett(at)Oracle-BI&EPM
    Play eSeminar Duration: 18 Minutes         Description: This webinar presents to ISV Partners Oracle’s latest release of BI Publisher, and describes how this tool can make their applications more competitive and appealing to their customers by providing High Quality Reporting and Business Intelligence embedded into their solution. • BI Publisher can Provide All Reports… at Lower Cost • Easier, with Better Developer Productivity • Better Managed : Better Performance, Less Administration • Highest Quality : Pixel Perfect and Interactive Reporting. Play eSeminar (Only accessible to Oracle Partners).

    Read the article

  • Google I/O 2012 - What's Next for Chrome Extensions?

    Google I/O 2012 - What's Next for Chrome Extensions? Mike West Chrome's extension system offers developers more opportunities than ever to customize and enhance users' experience on the web. New APIs are landing on a regular basis that provide new functionality and deeper hooks into Chrome itself. Join us for a walk through bleeding edge changes to Chrome's extension framework that increase security, improve performance, and make it easier than ever for users to get up and running with the product of your hard work. For all I/O 2012 sessions, go to developers.google.com From: GoogleDevelopers Views: 1372 29 ratings Time: 51:36 More in Science & Technology

    Read the article

  • Grub errors dual boot Windows 8 / Ubuntu 12.10

    - by luca-mastro
    I have got a newly bought ASUS N56V with Windows 8 preinstalled. I needed to install Ubuntu so i partitioned the disk and after having disabled the Secure Boot option from Windows 8 i successfully installed Ubuntu 12.10 from a Live USB. The problem is that if i try to access both to Windows 8 (loader) and Windows Recovery System (loader) from the grub menu, these tow errors show: "can't find command 'drivemap' " and "invalid EFI file path" and it goes back to the grub menu. In conclusion I do not have access to my Windows 8 partition but can only use Ubuntu. How can I solve the problem? I am pretty new to the matter. Thank you!

    Read the article

  • Implementing a Custom Coherence PartitionAssignmentStrategy

    - by jpurdy
    A recent A-Team engagement required the development of a custom PartitionAssignmentStrategy (PAS). By way of background, a PAS is an implementation of a Java interface that controls how a Coherence partitioned cache service assigns partitions (primary and backup copies) across the available set of storage-enabled members. While seemingly straightforward, this is actually a very difficult problem to solve. Traditionally, Coherence used a distributed algorithm spread across the cache servers (and as of Coherence 3.7, this is still the default implementation). With the introduction of the PAS interface, the model of operation was changed so that the logic would run solely in the cache service senior member. Obviously, this makes the development of a custom PAS vastly less complex, and in practice does not introduce a significant single point of failure/bottleneck. Note that Coherence ships with a default PAS implementation but it is not used by default. Further, custom PAS implementations are uncommon (this engagement was the first custom implementation that we know of). The particular implementation mentioned above also faced challenges related to managing multiple backup copies but that won't be discussed here. There were a few challenges that arose during design and implementation: Naive algorithms had an unreasonable upper bound of computational cost. There was significant complexity associated with configurations where the member count varied significantly between physical machines. Most of the complexity of a PAS is related to rebalancing, not initial assignment (which is usually fairly simple). A custom PAS may need to solve several problems simultaneously, such as: Ensuring that each member has a similar number of primary and backup partitions (e.g. each member has the same number of primary and backup partitions) Ensuring that each member carries similar responsibility (e.g. the most heavily loaded member has no more than one partition more than the least loaded). Ensuring that each partition is on the same member as a corresponding local resource (e.g. for applications that use partitioning across message queues, to ensure that each partition is collocated with its corresponding message queue). Ensuring that a given member holds no more than a given number of partitions (e.g. no member has more than 10 partitions) Ensuring that backups are placed far enough away from the primaries (e.g. on a different physical machine or a different blade enclosure) Achieving the above goals while ensuring that partition movement is minimized. These objectives can be even more complicated when the topology of the cluster is irregular. For example, if multiple cluster members may exist on each physical machine, then clearly the possibility exists that at certain points (e.g. following a member failure), the number of members on each machine may vary, in certain cases significantly so. Consider the case where there are three physical machines, with 3, 3 and 9 members each (respectively). This introduces complexity since the backups for the 9 members on the the largest machine must be spread across the other 6 members (to ensure placement on different physical machines), preventing an even distribution. For any given problem like this, there are usually reasonable compromises available, but the key point is that objectives may conflict under extreme (but not at all unlikely) circumstances. The most obvious general purpose partition assignment algorithm (possibly the only general purpose one) is to define a scoring function for a given mapping of partitions to members, and then apply that function to each possible permutation, selecting the most optimal permutation. This would result in N! (factorial) evaluations of the scoring function. This is clearly impractical for all but the smallest values of N (e.g. a partition count in the single digits). It's difficult to prove that more efficient general purpose algorithms don't exist, but the key take away from this is that algorithms will tend to either have exorbitant worst case performance or may fail to find optimal solutions (or both) -- it is very important to be able to show that worst case performance is acceptable. This quickly leads to the conclusion that the problem must be further constrained, perhaps by limiting functionality or by using domain-specific optimizations. Unfortunately, it can be very difficult to design these more focused algorithms. In the specific case mentioned, we constrained the solution space to very small clusters (in terms of machine count) with small partition counts and supported exactly two backup copies, and accepted the fact that partition movement could potentially be significant (preferring to solve that issue through brute force). We then used the out-of-the-box PAS implementation as a fallback, delegating to it for configurations that were not supported by our algorithm. Our experience was that the PAS interface is quite usable, but there are intrinsic challenges to designing PAS implementations that should be very carefully evaluated before committing to that approach.

    Read the article

  • Oracle Fusion Distributed Order Orchestration

    Designed from the ground-up using the latest technology advances and incorporating the best practices gathered from Oracle's thousands of customers, Fusion Applications are 100 percent open standards-based business applications that set a new standard for the way we innovate, work and adopt technology. Delivered as a complete suite of modular applications, Fusion Applications work with your existing portfolio to evolve your business to a new level of performance. In this AppCast, part of a special series on Fusion Applications, you hear lean how Oracle Fusion Distributed Order Orchestration can help companies improve customer service, reduce fulfillment costs, and optimize fulfillment decision making. Supporting a strategy for improving operational efficiency and boosting customer satisfaction, Fusion Distributed Order Orchestration alleviates or tempers critical production challenges many organizations face today by consolidating order information into a central location. You'll also discover how Fusion Distributed Order Orchestration works with your existing order management solutions.

    Read the article

  • Microsoft dévoile la version CTP de Dryad et DryadLINQ pour le développement parallèle et distribué

    Microsoft dévoile la version CTP de Dryad et DryadLINQ Pour le développement parallèle et distribué Microsoft vient de publier la version CTP de ses environnements de calculs parallèles et distribués qui seront commercialisés dans les semaines à venir, Dryad et DryadLINQ. Dryad est un moteur de calcul haute performance pour des calculs distribués, conçu pour simplifier la mise en oeuvre d'applications distribuées. DryadLINQ, lui, permet aux développeurs d'implémenter des applications Dryad en code managé à l'aide d'une version étendue du modèle de programmation LINQ. A la base, la technologie Dryad était un projet de recherche de Microsoft pour l'exécution des donnée...

    Read the article

  • Ubuntu hanging problem

    - by JMearsXS
    I have a problem with ubuntu 12.04 LTS on my custom tower pc in which the OS hangs at specific points:- - live usb with persistence: hangs when mouse appears. - installed to hard disk: hangs after login, logon sound never plays. i am a relative newbie with linux so any terminal output would not make sense to me :) the pc specs are as follows: - MSI MS-7301 Cuba motherboard - 3GB DDR2 RAM (NOT dual channel) - ATI Radeon HD 5450 gfx card the live usb works fine on all other PC's i tried it on. could any one help me ?

    Read the article

  • assistance recovering/reinstalling/installing ubuntu and win7

    - by razzrat
    New computer with Windows 7 installed, I defrag, shrink, re-boot from Ubuntu LiveUSB, go to gparted and look at partitions before installing Ubuntu....for some reason Win7 is still taking up 400G of my HD! I resized partition down with gparted, exit and yes of course I can't boot into Windows. When I go to install Ubuntu in new large unallocated space I get a blank screen at the point you are asked what kind of installation you want. I have Ubuntu 12.04 LiveUSB, Windows 7 re-installation disk and driver disks also. The HDD currently has 3 allocated partitions: 'diag' fat16, 'recovery' ntfs and 'OS' ntfs which has a red '!' next to it.

    Read the article

  • Google I/O 2010 - Writing real-time games for Android redux

    Google I/O 2010 - Writing real-time games for Android redux Google I/O 2010 - Writing real-time games for Android redux Android 201 Chris Pruett This session is a crash course in Android game development: everything you need to know to get started writing 2D and 3D games, as well as tips, tricks, and benchmarks to help your code reach optimal performance. In addition, we'll discuss hot topics related to game development, including hardware differences across devices, using C++ to write Android games, and the traits of the most popular games on Market. For all I/O 2010 sessions, please go to code.google.com From: GoogleDevelopers Views: 10 0 ratings Time: 58:57 More in Science & Technology

    Read the article

  • Wireless Very slow with 13.10 and BCM4313

    - by RyanCheu
    I have a laptop with the BCM4313, and it was working perfectly in Ubuntu 13.04, but I recently upgraded to 13.10 and now my wireless performance is horrible. Initially it didn't work at all, but I removed the wl driver and use the brcmsmac driver instead. Now when I boot up it works at the start, but gets progressively slower. My Android device is reporting 10Mbps down/20Mbps up, but my laptop only gets 1up/down. Does anyone know a solution? I really need my wireless to work, is my best option to just reinstall on 13.04? Thanks!

    Read the article

  • Nevron .NET Vision 2010 Vol.1 SP9 released VS2010 support

    We are pleased to announce the release of the new service pack for Nevron .NET Vision 2010 Vol.1. The service pack introduces: - Support for Visual Studio 2010 - new installation for VS2010 is available for download - Update for all installation version Nevron .NET Vision for VS2005, VS2008 and VS2010 - Significant performance improvements in the Chart Sampled Series Nevron .NET Vision 2010 Vol.1 SP9 is now available for download: http://www.nevron.com/Downloads.Download.aspx To download the latest...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • SQL Server 2008 R2: StreamInsight changes at RTM: Event Flow Debugger and Management Interface Secur

    - by Greg Low
    In CTP3, I found setting up the StreamInsight Event Flow Debugger fairly easy. For RTM, a number of security changes were made. First config: To be able to connect to the management interface, your user must be added to the Performance Log Users group. After you make this change, you must log off and log back on as the token is only added to your login token when you log on. I forgot this and spent ages trying to work out why I couldn't connect. Second config: You need to reserve the URL that the...(read more)

    Read the article

  • Efficient Way to Draw Grids in XNA

    - by sm81095
    So I am working on a game right now, using Monogame as my framework, and it has come time to render my world. My world is made up of a grid (think Terraria but top-down instead of from the side), and it has multiple layers of grids in a single world. Knowing how inefficient it is to call SpriteBatch.Draw() a lot of times, I tried to implement a system where the tile would only be drawn if it wasn't hidden by the layers above it. The problem is, I'm getting worse performance by checking if it's hidden than when I just let everything draw even if it's not visible. So my question is: how to I efficiently check if a tile is hidden to cut down on the draw() calls? Here is my draw code for a single layer, drawing floors, and then the tiles (which act like walls): public void Draw(GameTime gameTime) { int drawAmt = 0; int width = Tile.TILE_DIM; int startX = (int)_parent.XOffset; int startY = (int)_parent.YOffset; //Gets the starting tiles and the dimensions to draw tiles, so only onscreen tiles are drawn, allowing for the drawing of large worlds int tileDrawWidth = ((CIGame.Instance.Graphics.PreferredBackBufferWidth / width) + 4); int tileDrawHeight = ((CIGame.Instance.Graphics.PreferredBackBufferHeight / width) + 4); int tileStartX = (int)MathHelper.Clamp((-startX / width) - 2, 0, this.Width); int tileStartY = (int)MathHelper.Clamp((-startY / width) - 2, 0, this.Height); #region Draw Floors and Tiles CIGame.Instance.GraphicsDevice.SetRenderTarget(_worldTarget); CIGame.Instance.GraphicsDevice.Clear(Color.Black); CIGame.Instance.SpriteBatch.Begin(); //Draw floors for (int x = tileStartX; x < (int)MathHelper.Clamp(tileStartX + tileDrawWidth, 0, this.Width); x++) { for (int y = tileStartY; y < (int)MathHelper.Clamp(tileStartY + tileDrawHeight, 0, this.Height); y++) { //Check if this tile is hidden by layer above it bool visible = true; for (int i = this.LayerNumber; i <= _parent.ActiveLayer; i++) { if (this.LayerNumber != (_parent.Layers - 1) && (_parent.GetTileAt(x, y, i + 1).Opacity >= 1.0f || _parent.GetFloorAt(x, y, i + 1).Opacity >= 1.0f)) { visible = false; break; } } //Only draw if visible under the tile above it if (visible && this.GetTileAt(x, y).Opacity < 1.0f) { Texture2D tex = WorldTextureManager.GetFloorTexture((Floor)_floors[x, y]); Rectangle source = WorldTextureManager.GetSourceForIndex(((Floor)_floors[x, y]).GetTextureIndexFromSurroundings(x, y, this), tex); Rectangle draw = new Rectangle(startX + x * width, startY + y * width, width, width); CIGame.Instance.SpriteBatch.Draw(tex, draw, source, Color.White * ((Floor)_floors[x, y]).Opacity); drawAmt++; } } } //Draw tiles for (int x = tileStartX; x < (int)MathHelper.Clamp(tileStartX + tileDrawWidth, 0, this.Width); x++) { for (int y = tileStartY; y < (int)MathHelper.Clamp(tileStartY + tileDrawHeight, 0, this.Height); y++) { //Check if this tile is hidden by layers above it bool visible = true; for (int i = this.LayerNumber; i <= _parent.ActiveLayer; i++) { if (this.LayerNumber != (_parent.Layers - 1) && (_parent.GetTileAt(x, y, i + 1).Opacity >= 1.0f || _parent.GetFloorAt(x, y, i + 1).Opacity >= 1.0f)) { visible = false; break; } } if (visible) { Texture2D tex = WorldTextureManager.GetTileTexture((Tile)_tiles[x, y]); Rectangle source = WorldTextureManager.GetSourceForIndex(((Tile)_tiles[x, y]).GetTextureIndexFromSurroundings(x, y, this), tex); Rectangle draw = new Rectangle(startX + x * width, startY + y * width, width, width); CIGame.Instance.SpriteBatch.Draw(tex, draw, source, Color.White * ((Tile)_tiles[x, y]).Opacity); drawAmt++; } } } CIGame.Instance.SpriteBatch.End(); Console.WriteLine(drawAmt); CIGame.Instance.GraphicsDevice.SetRenderTarget(null); //TODO: Change to new rendertarget instead of null #endregion } So I was wondering if this is an efficient way, but I'm going about it wrongly, or if there is a different, more efficient way to check if the tiles are hidden. EDIT: For example of how much it affects performance: using a world with three layers, allowing everything to draw no matter what gives me 60FPS, but checking if its visible with all of the layers above it gives me only 20FPS, while checking only the layer immediately above it gives me a fluctuating FPS between 30 and 40FPS.

    Read the article

  • Does data size in TCP/UDP make a difference on transmission time

    - by liortal
    While discussing the development of a network component for our game engine, a member of our team suggested that transmitting either 500 bytes or 1k of data using UDP makes no difference from performance perspective of the system (the time it takes to transmit the data). He actually said that as long as you don't cross the MTU size, the size of the transmitted data doesn't really matter as it's all the same. Is that true for UDP? what about TCP? That sounds just plain wrong to me, but i am not a network expert. *I've been reading about other companies' game networking architectures, and it seems they're all trying to keep transmitted data to a minimum, making my colleague's claims seem even more unreasonable.

    Read the article

  • Spezialisierung ohne Grenzen

    - by A&C Redaktion
    Arrow erreicht Exadata Spezialisierung für alle EMEA-Länder “Know-how sells” – das weiß auch unser VAD Arrow. Der IT-Distributor aus Fürstenfeldbruck, nahe München, hat sich auf die Bereitstellung von Enterprise und Midrange Computing Lösungen fokussiert. So auch für die Exadata Technologie von Oracle. Exadata beinhaltet Server, Speicher, Netzwerktechnik und Datenbanksoftware in einem System und hilft so, auch große Datenmengen – die „Big Data“ – spielend zu managen. Die Kombination aus Hard- und Software bietet Oracle Partnern enorme Geschäftspotenziale im Verkauf und im Service, deshalb ist eine Expertise so wichtig. Durch die vier europäischen Demo-Zentren und insgesamt acht komplett installierte Exadata reichlich Erfahrung mit der Oracle Exa-Familie sammeln können. Der VAD bietet Oracle Partnern und Kunden Performance-Tests, Testumgebungen und Proof of Concepts (PoC) an – und das länderübergreifend. Als logische Konsequenz wurde Arrow im August 2012 mit der EMEA Spezialisierung für Exadata von Oracle ausgezeichnet! Wir gratulieren ganz herzlich und wünschen viel Erfolg mit dem Exa-Stack!

    Read the article

  • Spezialisierung ohne Grenzen

    - by A&C Redaktion
    Arrow erreicht Exadata Spezialisierung für alle EMEA-Länder “Know-how sells” – das weiß auch unser VAD Arrow. Der IT-Distributor aus Fürstenfeldbruck, nahe München, hat sich auf die Bereitstellung von Enterprise und Midrange Computing Lösungen fokussiert. So auch für die Exadata Technologie von Oracle. Exadata beinhaltet Server, Speicher, Netzwerktechnik und Datenbanksoftware in einem System und hilft so, auch große Datenmengen – die „Big Data“ – spielend zu managen. Die Kombination aus Hard- und Software bietet Oracle Partnern enorme Geschäftspotenziale im Verkauf und im Service, deshalb ist eine Expertise so wichtig. Durch die vier europäischen Demo-Zentren und insgesamt acht komplett installierte Exadata reichlich Erfahrung mit der Oracle Exa-Familie sammeln können. Der VAD bietet Oracle Partnern und Kunden Performance-Tests, Testumgebungen und Proof of Concepts (PoC) an – und das länderübergreifend. Als logische Konsequenz wurde Arrow im August 2012 mit der EMEA Spezialisierung für Exadata von Oracle ausgezeichnet! Wir gratulieren ganz herzlich und wünschen viel Erfolg mit dem Exa-Stack!

    Read the article

  • Data base preference for network based C# windows application [on hold]

    - by Sinoop Joy
    I'm planning to develop a C# widows based application for an academy. The academy will have different instances of application running in different machines. The database should have shared access. All the application instances can do update, delete or insert. I've not done any network based application. Anybody can give any useful link to where to start with ? Which database would give max performance with all required features i said for this scenario ?

    Read the article

  • Updated copy of the OBIEE Tuning whitepaper

    - by inowodwo
    The Product Assurance team have released an updated copy of the OBIEE Tuning Whitepaper. You can find it on the PA blog https://blogs.oracle.com/pa/entry/test or via Support note OBIEE 11g Infrastructure Performance Tuning Guide (Doc ID 1333049.1) https://support.us.oracle.com/oip/faces/secure/km/DocumentDisplay.jspx?id=1333049.1&recomm=Y This new revised document contains following useful tuning items: 1.    New improved HTTP Server caching algorithm. 2.    Oracle iPlanet Web Server tuning parameters. 3.    New tuning parameters settings / values for OPIS/OBIS components.

    Read the article

< Previous Page | 482 483 484 485 486 487 488 489 490 491 492 493  | Next Page >