Search Results

Search found 22961 results on 919 pages for 'memory management'.

Page 98/919 | < Previous Page | 94 95 96 97 98 99 100 101 102 103 104 105  | Next Page >

  • Oracle Utilities Mobile Workforce Management V2.0 has arrived

    - by Anthony Shorten
    it is finally upon us. Oracle Utilities Mobile Workforce Management (MWM) V2.0 has been released and is now available (see Press Release). This is significant for me as this is the first product to use the new version of the Oracle Utilities Application Framework V4.0.1. This release is very significant as it adds a lot of new functionality to the framework, not just for MWM but will progressively rolled out across a few moew Oracle Utilities products over the next 12 months. Watch the skies for more annoucements. Now that Framework 4.0.1 has been released I will be updating this blog ona regular basis outlining significant features (there are over 60+ features in the new Framework) for you too understand. It has been hard work but it finally has been released and used by the first product off the assembly line we call product development.

    Read the article

  • Video White Paper: Mega-Project Management: Reducing Risk & Complexity across the Value Chain

    - by Melissa Centurio Lopes
    Watch this short video white paper, to learn how Oracle Primavera can help you keep projects on track and protect your investments. You can also download the full white paper “Mega-Project Management: Reducing Risk & Complexity Across the Value Chain” to gain more in depth information about strategies for collaborating and sharing information and data in a systematic way across the value chain. Download the white paper in order to learn how your company can get the expected payoff from your next mega project. Register now to download the full complementary white paper, and discover how to: Improve decision-making and accountability through enterprise-wide visibility, workflows, and collaboration Reduce financial and performance risk

    Read the article

  • mp3 file streaming/download - apache server memory issue

    - by Manolis
    I have a website, in which users can upload mp3 files (uploadify), stream them using an html5 player (jplayer) and download them using a php script (www.zubrag.com/scripts/). When a user uploads a song, the path to the audio file is saved in the database and i'm using that data in order to play and show a download link for the song. The problem that i'm experiencing is that, according to my host, this method is using a lot of memory on the server, which is dedicated. Link to script: http://pastebin.com/Vus8SRa7 How should I handle the script properly? And what would be the best way to track down the problem? Any ideas on cleaning up the code? Any help much appreciated.

    Read the article

  • Oracle Transportation Management Annual Customer Conference

    - by [email protected]
    The 2010 Oracle Transportation Management (OTM) Conference will be held June 13-16 in Philadelphia, Pennsylvania. The conference brings together all things OTM: users, prospective users, development personnel, product strategy, implementation experts, and software and services partners.  With over 200 attendees, this conference is the premiere location and time to learn about OTM, build relationships with peers, and get answers to all your OTM questions.    This year's conference will be held at the: Sheraton Society Hill, One Dock St., Philadelphia, PA. 19106Companies speaking at this year's event include:AT&T Land O Lakes BlueScope Steel Baillie Lumber Kraft Sears Roseburg Forest Products Toyota Beckman CoulterLevi StraussNiagara BottlingSmurfit StonePQ CorporationOffice Depot               To register click here http://www.otmconference.com/ConfAgenda.aspx.  1

    Read the article

  • Why does one loop take longer to detect a shared memory update than another loop?

    - by Joseph Garvin
    I've written a 'server' program that writes to shared memory, and a client program that reads from the memory. The server has different 'channels' that it can be writing to, which are just different linked lists that it's appending items too. The client is interested in some of the linked lists, and wants to read every node that's added to those lists as it comes in, with the minimum latency possible. I have 2 approaches for the client: For each linked list, the client keeps a 'bookmark' pointer to keep its place within the linked list. It round robins the linked lists, iterating through all of them over and over (it loops forever), moving each bookmark one node forward each time if it can. Whether it can is determined by the value of a 'next' member of the node. If it's non-null, then jumping to the next node is safe (the server switches it from null to non-null atomically). This approach works OK, but if there are a lot of lists to iterate over, and only a few of them are receiving updates, the latency gets bad. The server gives each list a unique ID. Each time the server appends an item to a list, it also appends the ID number of the list to a master 'update list'. The client only keeps one bookmark, a bookmark into the update list. It endlessly checks if the bookmark's next pointer is non-null ( while(node->next_ == NULL) {} ), if so moves ahead, reads the ID given, and then processes the new node on the linked list that has that ID. This, in theory, should handle large numbers of lists much better, because the client doesn't have to iterate over all of them each time. When I benchmarked the latency of both approaches (using gettimeofday), to my surprise #2 was terrible. The first approach, for a small number of linked lists, would often be under 20us of latency. The second approach would have small spats of low latencies but often be between 4,000-7,000us! Through inserting gettimeofday's here and there, I've determined that all of the added latency in approach #2 is spent in the loop repeatedly checking if the next pointer is non-null. This is puzzling to me; it's as if the change in one process is taking longer to 'publish' to the second process with the second approach. I assume there's some sort of cache interaction going on I don't understand. What's going on?

    Read the article

  • Leveraging AutoVue in Oracle's Universal Content Management for Improved Document

    AutoVue visualization, leveraged within Oracle’s Universal Content Management, makes access to technical information widely available to UCM users, allowing them to review and collaborate on CAD and engineering content in a variety of business processes and workflows. Comments and feedback are captured within the design context and recorded and tracked digitally within UCM, providing a reliable trail of decisions and approvals thereby facilitating an organization’s audit compliance. The joint solution can also be leveraged in broader Oracle applications, such as Web Center, eAM to name a few. Hear about the benefits UCM users can achieve by introducing AutoVue visualization into their UCM environment.

    Read the article

  • How can i estimate memory usage of stl::map?

    - by Drakosha
    For example, I have a std::map with known sizeof(A) and sizefo(B), while map has N entries inside. How would you estimate its memory usage? I'd say it's something like (sizeof(A) + sizeof(B)) * N * factor But what is the factor? Different formula maybe? Update: Maybe it's easier to ask for upper bound?

    Read the article

  • Updated article "Agent Alerts Management Pack"

    - by TiborKaraszi
    I've just updated the "Agent Alerts Management Pack" found here . I realize that some don't feel confident in reading and executing T-SQL code and they instead prefer to point & click in SSMS instead. So I added two tables with my suggestion on the severity levels and error numbers to define alerts for. Share this post: email it! | bookmark it! | digg it! | reddit! | kick it! | live it!...(read more)

    Read the article

  • Oracle Applications Day 2012 -Experience the Global Innovation of Management Applications

    - by antonella.buonagurio
    Il 10 ottobre a Milano e il 17 ottobre a Roma si sono svolti gli Oracle Applications Day, dedicati alla community di Clienti e Partner Oracle. Le due giornate hanno visto la partecipazione di più di 400 persone che hanno condiviso le loro esperienze e conoscenze in ambito applicativo. Durante la sessione plenaria sono state illustrate tutte le novità relative alle Oracle Applications ed in particolare le Oracle Fusion Applications mentre durante le 2 giornate più di 20 clienti hanno parlato di come utilizzano in modo strategico e con successo le soluzioni Oracle. 15 Business Partner grazie all'iniziativa "Partner Instant Workshop" hanno incontrato direttamente i clienti e discusso delle tematiche più calde del momento. Se non hai potuto partecipare all'evento oppure vuoi rivivere quei momenti qui sotto trovi la presentazione della plenaria mentre cliccando su ciascun titolo delle sessioni parallele puoi trovare le rispettive presentazioni. Innovation for Human Resources Performance Management Excellence Empower Applications with Technology (tenutasi solo a Milano) Applications for Public Sector (tenutasi solo a Roma) Next Generation Global Operations Customer Experience Revolution

    Read the article

  • Oracle ETPM is renamed Oracle Public Sector Revenue Management (PSRM)

    - by Rick Finley
    We are excited to announce to that with the upcoming release of v2.4, we are renaming ETPM to Oracle Public Sector Revenue Management (Oracle PSRM).  This is a pure name change, and all terms and conditions for existing customer licensing remain unchanged.  We feel that this updated naming is a better reflection of our current customer base, which includes tax revenue for many Departments of Revenue, as well as agencies that at manage non-tax revenue, such as regulatory fees, loans, and social benefits.    Please note, as part of this name change, related products in the Oracle ETPM family, such as Oracle Tax Analytics, and Oracle ETPM Self Service, will be renamed at their next major product release to align to the Oracle PSRM theme.   

    Read the article

  • cocos2d-x - object creation and management in game design

    - by Jason
    How do others keep track of everything going on in their games? I am working on a new game and I am quickly realizing everything that I need to keep track of. Example: Maybe a layerManager that keeps track of all the layers and what is happening for a particular scene. Maybe a sceneManager for sharing objects among scenes But then getting to game play itself, what if you have 100 objects on the screen each with its own state and happenings, there needs tobe a way to keep track of all of that. Drawing everything out is really helping me. Can anyone share with me how they go about object tracking/management? I am seeing a few different managers and then maybe even a parent object that manages the managers..is my thinking way off? Any design patterns that may be useful for me to read about? Update: doing some reading and maybe a Factory pattern might apply.

    Read the article

  • Creating a remote management interface

    - by Johnny Mopp
    I'm looking for info on creating a remote management interface for our software. This is not anything illicit. Our software is for live TV production and once they go on-air we can't access the PC (usually through LogMeIn). I would like to be able to upload/download files and issue commands to our software. The commands would be software specific like "load this file" or "run this script" or "return this value" etc. A socket connection is preferred but the problem is most of our PCs are behind firewalls and NAT servers. I'm not sure where to start. I think HTTP tunneling is the way to go but am wondering if there are other options or recommendations. Also, assume our clients are not willing to open up ports for security reasons. Thanks.

    Read the article

  • Book review: SQL Server Transaction Log Management

    - by Hugo Kornelis
    It was an offer I could not resist. I was promised a free copy of one of the newest books from Red Gate Books , SQL Server Transaction Log Management (by Tony Davis and Gail Shaw ), with the caveat that I should write a review after reading it. Mind you, not a commercial, “make sure we sell more copies” kind of review, but a review of my actual thoughts. Yes, I got explicit permission to be my usual brutally honest self. A total win/win for me! First, I get a free book – and free is always good,...(read more)

    Read the article

  • Dell whitepaper on PowerEdge R810 R910 and M910 Memory Architecture

    - by jchang
    The Dell PowerEdge 11 th Generation Servers: R810, R910 and M910 Memory Guidance whitepaper seems to have caused some confusion. I believe the source is an error in the paper. In the section on FlexMem Bridge Technology, the Dell whitepaper says this applies to the R810 and the M910. The Dell M910 is a 4-way blade server for the Xeon 7500 series processor line. First a breif recap. The R810 is a 2-way server, by which I mean it has two sockets regardless of the number of cores on each processor....(read more)

    Read the article

  • recurring billing / profiles management system

    - by Karl Cassar
    As a company, we have various recurring fees which our clients pay - these can include: hosting plans maintenance agreements SLAs ... I would like to know if anyone knows of a good, web-based recurring billing / payments management system which we could use to help us get more organised regarding this aspect of our business. Basically, we would need to: Create recurring profiles, e.g: Hosting, emails / domain services @ 200eur / year Be able to give free / extend the subscription period, for any reason. Also, we don't have specific products which we would like to choose and charge - all these recurring fees are discussed with the clients, and are created on a per-client basis. I'm not sure if this is the best place to ask for, however since I think most 'webmasters' require such a system to keep track of payment, I thought this would be the place to go for. Thanks in advance!

    Read the article

  • Oracle Database In-Memory: Launch in Frankfurt

    - by Carsten Czarski
    Diesmal gibt es etwas Altes ... und etwas Neues. Zuerst das Neue: Am 11. Juni wird Larry Ellison in Redwood Shores die neue, bahnbrechende Oracle Database In-Memory Funktionalität vorstellen. Mit dieser neuen Technologie profitieren Kunden von beschleunigter Datenbankleistung für Analytics, Data Warehousing, Reporting und Online Transaction Processing (OLTP). Nur 6 Tage später - am 17. Juni -  findet, in Frankfurt, der einzige europäische Launch-Event statt. Neben Fachvorträgen, Panelveranstaltung und Demos wird ein Vortrag von Andy Mendelsohn, Head of Database Product Development, vorgesehen. Melden Sie sich heute noch an. Und hier ist das Alte: Wer erinnert sich noch die die HTML DB ...? In den Archiven der APEX Community Seite haben wir ein Video gefunden, welches zeigt, wie man Seiten in der HTML DB für andere Entwickler sperren konnte. Das gibt es heute übrigens auch noch - es sieht nur etwas anders aus. Viel Spaß beim Ansehen.

    Read the article

  • New Oracle Information Rights Management release (11.1.1.3)

    - by Simon Thorpe
    Just released is the latest version of the market leading document security technology from Oracle. Oracle IRM 11g is the result of over 12 years of development and innovation to allow customers to provide persistent security to their most confidential documents and emails. This latest release continues our refinement of the technology and features the following; Continued improvements to the web based Oracle IRM Management Website New features in the out of the box classification model New Java APIs improving application integration support Support for DB2 as the IRM database. Over the coming months we will see more releases from this technology as we improve format support, platform support and continue the strategy to for Oracle IRM as the most secure, scalable and usable document security solution in the market. Want to learn more about Oracle IRM? View our video presentation and demonstration or try using it for your self via our simple online self service demo. Keep up to date on Oracle via this blog or on our Twitter, YouTube and Facebook pages.

    Read the article

  • In-Memory OLTP Sample for SQL Server 2014 RTM

    - by Damian
    I have just found a very good resource about Hekaton (In-memory OLTP feature in the SQL Server 2014). On the Codeplex site you can find the newest Hekaton samples - https://msftdbprodsamples.codeplex.com/releases/view/114491. The latest samples we have were related to the CTP2 version but the newest will work with the RTM version.There are some issues fixed you might find if you tried to run the previous samples on the RTM version:Update (Apr 28, 2014): Fixed an issue where the isolation level for sample stored procedures demonstrating integrity checks was too low. The transaction isolation level for the following stored procedures was updated: Sales.uspInsertSpecialOfferProductinmem, Sales.uspDeleteSpecialOfferinmem, Production.uspInsertProductinmem, and Production.uspDeleteProductinmem. 

    Read the article

  • How to I serialize a large graph of .NET object into a SQL Server BLOB without creating a large bu

    - by Ian Ringrose
    We have code like: ms = New IO.MemoryStream bin = New System.Runtime.Serialization.Formatters.Binary.BinaryFormatter bin.Serialize(ms, largeGraphOfObjects) dataToSaveToDatabase = ms.ToArray() // put dataToSaveToDatabase in a Sql server BLOB But the memory steam allocates a large buffer from the large memory heap that is giving us problems. So how can we stream the data without needing enough free memory to hold the serialized objects. I am looking for a way to get a Stream from SQL server that can then be passed to bin.Serialize() so avoiding keeping all the data in my processes memory. Likewise for reading the data back... Some more background. This is part of a complex numerical processing system that processes data in near real time looking for equipment problems etc, the serialization is done to allow a restart when there is a problem with data quality from a data feed etc. (We store the data feeds and can rerun them after the operator has edited out bad values.) Therefore we serialize the object a lot more often then we de-serialize them. The objects we are serializing include very large arrays mostly of doubles as well as a lot of small “more normal” objects. We are pushing the memory limit on a 32 bit system and make the garage collector work very hard. (Effects are being made elsewhere in the system to improve this, e.g. reusing large arrays rather then create new arrays.) Often the serialization of the state is the last straw that courses an out of memory exception; our peak memory usage is while this serialization is being done. I think we get large memory pool fragmentation when we de-serialize the object, I expect there are also other problem with large memory pool fragmentation given the size of the arrays. (This has not yet been investigated, as the person that first looked at this is a numerical processing expert, not a memory management expert.) Are customers use a mix of Sql Server 2000, 2005 and 2008 and we would rather not have different code paths for each version of Sql Server if possible. We can have many active models at a time (in different process, across many machines), each model can have many saved states. Hence the saved state is stored in a database blob rather then a file. As the spread of saving the state is important, I would rather not serialize the object to a file, and then put the file in a BLOB one block at a time. Other related questions I have asked How to Stream data from/to SQL Server BLOB fields? Is there a SqlFileStream like class that works with Sql Server 2005?

    Read the article

  • Performance-Based Management Stinks

    - by andyleonard
    Introduction This post is the forty-eighth part of a ramble-rant about the software business. The current posts in this series can be found on the series landing page . This post is about Performance-Based Management (PBM). Almost… In Mere Christianity , C. S. Lewis refutes an argument with the following statement: It has every amiable quality except that of being useful. I feel the same way about PBM. I am a metrics person. I thrive – intellectually, emotionally, and economically – on business intelligence...(read more)

    Read the article

  • OpenGL Shading Program Object Memory Requirement

    - by Hans Wurst
    gDEbugger states that OpenGL's program objects only occupy an insignificant amount of memory. How much is this actually? I don't know if the stuff I looked up in mesa is actually that I was looking for but it requires 16KB [Edit: false, confusing struct names, less than 1KB immediate, some further behind pointers] per program object. Not quite insignificant. So is it recommended to create a unique program object for each object of the scene? Or to share a single program object and set the scene's object's custom variables just before its draw call?

    Read the article

  • Grails application hogging too much memory

    - by RN
    Tomcat 5.5.x and 6.0.x Grails 1.6.x Java 1.6.x OS CentOS 5.x (64bit) VPS Server with memory as 384M export JAVA_OPTS='-Xms128M -Xmx512M -XX:MaxPermSize=1024m' I have created a blank Grails application i.e simply by giving the command grails create-app and then WARed it I am running Tomcat on a VPS Server When I simply start the Tomcat server, with no apps deployed, the free memory is about 236M and used memory is about 156M When I deploy my "blank" application, the memory consumption spikes to 360M and finally the Tomcat instance is killed as soon as it takes up all free memory As you have seen, my app is as light as it can be. Not sure why the memory consumption is as high it is. I am actually troubleshooting a real application, but have narrowed down to this scenario which is easier to share and explain.

    Read the article

  • In-Memory OLTP Sample for SQL Server 2014 RTM

    - by Damian
    I have just found a very good resource about Hekaton (In-memory OLTP feature in the SQL Server 2014). On the Codeplex site you can find the newest Hekaton samples - https://msftdbprodsamples.codeplex.com/releases/view/114491. The latest samples we have were related to the CTP2 version but the newest will work with the RTM version.There are some issues fixed you might find if you tried to run the previous samples on the RTM version:Update (Apr 28, 2014): Fixed an issue where the isolation level for sample stored procedures demonstrating integrity checks was too low. The transaction isolation level for the following stored procedures was updated: Sales.uspInsertSpecialOfferProductinmem, Sales.uspDeleteSpecialOfferinmem, Production.uspInsertProductinmem, and Production.uspDeleteProductinmem. 

    Read the article

  • How can I give eclipse more memory than 512M?

    - by newbie
    I have following setup, but when I put 1024 and replace all 512 with 1024, then eclipse won't start at all. How can I have more than 512M memory for my eclipse JVM? -startup plugins/org.eclipse.equinox.launcher_1.0.201.R35x_v20090715.jar --launcher.library plugins/org.eclipse.equinox.launcher.win32.win32.x86_1.0.200.v20090519 -product com.springsource.sts.ide --launcher.XXMaxPermSize 512M -vm C:\Program Files (x86)\Java\jdk1.6.0_18\bin\javaw -vmargs -Dosgi.requiredJavaVersion=1.5 -Xms512m -Xmx512m -XX:MaxPermSize=512m

    Read the article

< Previous Page | 94 95 96 97 98 99 100 101 102 103 104 105  | Next Page >