Search Results

Search found 552 results on 23 pages for 'lady disaster'.

Page 11/23 | < Previous Page | 7 8 9 10 11 12 13 14 15 16 17 18  | Next Page >

  • Pain of the Week/Expert's Perspective: Performance Tuning for Backups and Restores

    - by KKline
    First off - the Pain of the Week webcast series has been renamed. It's now known as The Expert's Perspective . Please join us for future webcasts and, if you're interested in speaking, drop me a note to see if we can get you on the roster! The bigger your databases get, the longer backups take. That doesn't really seem like a huge problem — until disaster strikes and you need to restore your databases as fast as possible. Join my buddy Brent Ozar ( blog | twitter ), a Microsoft Certified Master of...(read more)

    Read the article

  • Why do large IT projects tend to fail or have big cost/schedule overruns?

    - by Pratik
    I always read about large scale transformation or integration project that are total or almost total disaster. Even if they somehow manage to succeed the cost and schedule blow out is enormous. What is the real reason behind large projects being more prone to failure. Can agile be used in these sort of projects or traditional approach is still the best. One example from Australia is the Queensland Payroll project where they changed test success criteria to deliver the project. See some more failed projects in this SO question Have you got any personal experience to share?

    Read the article

  • Printer "ripping" forever (network printer)

    - by Julien Gorenflot
    Since I installed Ubuntu 11.10, printing is a disaster. I did not have the problem with Lucid Lynx (Ubuntu 10.04), but maybe it just comes from the fact that someone else had installed it for me, and possibly it configured better. When I print a pdf, even 2 pages, my printer (SHARP MX 2300N) stays in rippen for hours. "Rippen" is a German word, not really sure how to translate. Google translate says, The English equivalent is "Rib". And eventually, sometimes, the pages finally get printed. But in between my whole floor is very angry because they also need the printer. Additionally, I don't always have the whole day for waiting for my pages. I remember that when printing I used to be asked if I wanted to reduce transparency effects, which does not seem to happen anymore after I installed Ubuntu 11.10. Is there any connection? Not sure, because I don't think it was for pdf files.

    Read the article

  • What issues carry the highest risk in a software project?

    - by Mehrdad
    Clearly, software projects are different from other industries in terms of many things like for instance, quality assurance, project progress measurement, and many other things. Unique characteristics of software projects also makes the risk management process unique. Lots of issues in a project might lead it to unacceptable delay or failure to deliver business value. They might even make a complete disaster in the project. What are the deadliest risk factors in a software project? How to analyze, prevent and handle them? Particularly, I'm interested in the issues that you can detect from the beginning and you should keep an eye on (for example, you might be told about a third-party API that the current application uses and lacks documentation). Please share your experiences if they are relevant.

    Read the article

  • Common practice in handling bounce message

    - by foodil
    At now I mainly create a mail account separately (with different domain name [email protected]) and i add this mail as one return path. So the bounce message will only go to that mailbox and i parse the mail message one by one to check the failure receipent and the error code, then i convert the error code to the actual error message. Finally, the error message and the fail receipent's mail are post to my system and let my system user check the bounce information. Is it a common practice? Since i am worry about the mail other from bounce message have sent to my mail box, that would be a disaster if i parse them without filter them out, but how can i filter out between bounce message and normal mail? Thank you for any kind of help.

    Read the article

  • What is the term for a 'decoy' feature or intentional bug?

    - by Freiheit
    I have forgotten a slang programming term. This thing is an intentional bug or a decoy feature used as a distraction. An example usage, "Hey Bob, QA is doing a review today. Put a $THING into the module so they actually have a problem to find". This can be used negatively, to have a very obvious intentional flaw to discover as a distraction from a real problem. This can also be used positively. Its like how you always let rescue dogs 'find' a victim when searching a disaster area. It can also be used to verify that a QA process is actually catching flaws. What is the term I am looking for?

    Read the article

  • Bringing in New Architecture During Maintenance on Legacy Systems

    - by Mike L.
    I have been tasked with adding some new features to a legacy ASP.NET MVC2 project. The codebase is a disaster and I want to write these new features with some thought behind the implementation and not just throw these new features into the mess. I would like to introduce things like dependency injection and the orchestrator pattern; just to the code that I am going to write. I don't have enough time to try to refactor the entire system. Is it OK to not be consistent with the rest of the codebase and add new features following different design principles? Should I not introduce new patterns and just get the features implemented? I feel like it might be confusing to the next person to see parts of the system using a design that other parts are not following.

    Read the article

  • What issues carry the highest risk in a software project?

    - by Mehrdad
    Clearly, software projects are different from other industries in terms of many things like for instance, quality assurance, project progress measurement, and many other things. Unique characteristics of software projects also makes the risk management process unique. Lots of issues in a project might lead it to unacceptable delay or failure to deliver business value. They might even make a complete disaster in the project. What are the deadliest risk factors in a software project? How to analyze, prevent and handle them? Particularly, I'm interested in the issues that you can detect from the beginning and you should keep an eye on (for example, you might be told about a third-party API that the current application uses and lacks documentation). Please share your experiences if they are relevant.

    Read the article

  • Oracle OpenWorld Update -- Highly Available WebLogic Messaging Architectures: Sharing a Customer Experience with Comcast

    - by Ruma Sanyal
    This session will describe a Comcast’s hands-on  experience using WebLogic JMS as their high-performance enterprise messaging system including high availability, and disaster recovery capabilities as Comcast is rolling out a cross-site active-active message bus. In the session, we will cover the following: Key capabilities in WebLogic JMS that enabled Comcast to design such an architecture Details of the architecture put in place Details about application design needed to make all of this successful Failover and fail back processes The results from this new architecture are higher availability, better performance, more flexibility, and reduced costs through better utilization of hardware and improved manageability. For more information about this and other WebLogic sessions, review the Oracle WebLogic Focus On document here. Details: Tuesday, Oct 2, 5-6pm, Moscone South Room 306

    Read the article

  • Webcast: Maximum Availability with Oracle GoldenGate

    - by Markus Weber
    Make sure to register here for this upcoming live webcast, with great speakers, and a live Q&A chat session.It is scheduled for Thursday, December 13th, 2012, at 10am Pacific Time. There are no doubts about the strength, and depth, of Oracle's offerings for highly available database environments. But depending on the nature of your data center environments, your different uses cases and workloads, your service level agreements and expectations, there might still be some open questions about what the best solutions and architectures are to assure high availability and/or affordable, efficient disaster recovery. Therefore, this webcast will answer questions such as: How does Oracle GoldenGate fit into Oracle's Maximum Availability Architecture ? When to use Oracle Active Data Guard ? When Oracle GoldenGate ? When both ? When Oracle GoldenGate is a definite requirement for Database High Availability ? Overall, what are new high availability features in Oracle GoldenGate 11g Release 2 ? 

    Read the article

  • SQLServerCentral Webinar Series #21 - Forgotten Rings & Other Monitoring Stories

    Most common monitoring metrics are important and useful, especially over time, but they can fall short. How do you gather information to determine, for example, if you have buffer cache pressure? Register now for the free webinar. Wednesday, October 17 2012 4:00pm - 5:00pm BST Are you sure you can restore your backups? Run full restore + DBCC CHECKDB quickly and easily with SQL Backup Pro's new automated verification. Check for corruption and prepare for when disaster strikes. Try it now.

    Read the article

  • Another Update to SQL Server Configuration Scripting Utility

    - by Bill Graziano
    I’ve been gradually adding features to my utility that scripts the configuration of a SQL Server.  Since my last post I’ve added the following features: Skip any encrypted object in a database Script alerts, alert notifications and operators Script audits Always script model, master and msdb to capture any user-defined objects in those databases Logins are now scripted so that everything for a login is grouped together. There’s a second section in the logins that handles default databases.  In many cases a login’s default database is a mirror target and can’t be set.  This is now handled gracefully.  It also includes a separate section for all default databases so those can be quickly set in the event of a disaster. Script credentials Script proxy accounts Script database mail My goal is still to get everything outside a database scripted.  This release is enough that I can keep my mirror target servers in sync with their principals.

    Read the article

  • SQL in the City - Seattle 2012

    Start the week in Seattle off with a free day of training on Nov 5, 2012 with SQL in the City. Grant Fritchey, Steve Jones and more will be talking SQL Server in the Pacific Northwest. Join us and debate and discuss SQL Server the Red Gate Way. Are you sure you can restore your backups? Run full restore + DBCC CHECKDB quickly and easily with SQL Backup Pro's new automated verification. Check for corruption and prepare for when disaster strikes. Try it now.

    Read the article

  • Should we design programs to randomly kill themselves?

    - by jimbojw
    In a nutshell, should we design death into our programs, processes, and threads at a low level, for the good of the overall system? Failures happen. Processes die. We plan for disaster and occasionally recover from it. But we rarely design and implement unpredictable program death. We hope that our services' uptimes are as long as we care to keep them running. A macro-example of this concept is Netflix's Chaos Monkey, which randomly terminates AWS instances in some scenarios. They claim that this has helped them discover problems and build more redundant systems. What I'm talking about is lower level. The idea is for traditionally long-running processes to randomly exit. This should force redundancy into the design and ultimately produce more resilient systems. Does this concept already have a name? Is it already being used in the industry?

    Read the article

  • How to convince the agile product owner to change their mind? [closed]

    - by Joshiatto
    A friend of a friend ran into a situation recently in which the agile product owner specified features down to exactly what every single user click should look like. The problem is, the dev team has already figured out a way to accomplish the business value in fewer clicks (better UX), but in the past, questioning the product owner has led to career disaster. How do we convince the product owner to change their mind and go along with our recommendation? What can be done within the agile model to fix this situation and how do we accomplish it? On a bigger level: What can be done to make agile product owners better at their job to prevent this kind of thing from happening?

    Read the article

  • Which is the best low/mid level GPU in terms of compatibility for 12.04 and 12.10?

    - by Ghost
    I'm tired of dealing with the disaster that is catalyst on a 4xxx IGP and I want to buy some discrete graphics that support both Unity and alternative environments like GNOME3 and Mate-Cinnamon without getting sluggish. Besides that, 1080p video and some light 3D gaming. Again nothing crazy, I'm not a designer and I wont be using stuff like Blender. One thing I want to make clear is that I want a GPU that is 100% compatible, that means no bugs, no glitches, not having to toggle between propietary and Xorg because it decided to crapout on me. Thanks for your help

    Read the article

  • Handling Backups for Rapid Resilience

    The backup and restore system in SQL Server hasn't changed a great deal over the years despite a huge growth in the typical size of databases. When disaster strikes, and an important service is taken offline while a restore is performed, there is often time to reflect on whether it might be possible to design databases for a more rapid recovery of the most critical parts of a database application. Uncover the flexibility of custom metricsKeep up to date from anywhere with SQL Monitor, and monitor the most important data for your servers and applications. Download a free trial.

    Read the article

  • SmartGWT alternatives

    - by Roger
    I have been using SmartGWT but have run into trouble mixing SmartGWT with other frameworks such as Wicket. Infact it has been a bit of a disaster and I regret going with it in the first place. I do not want this question to turn into any type of flame war, so therefore would just like people to list alternatives and their experience with those alternatives.

    Read the article

  • Binding Bitmapimge to Image in Wpf?

    - by Rev
    This is a simple Question (lets see) I want to bind bitmap image to Image. For doing this in cs code u must write this line. this.leftImage.Source = new BitmapImage(new Uri(@"C:\a.bmp")); But I want make Binding from resources. Because In release time resources became part of project.exe file and if you make binding from file(Mean set Image.source with Image file address), you must always put image file in the same address(disaster programing) :)

    Read the article

  • Why am I getting unexpected output trying to write a hash structure to a file?

    - by Harm De Weirdt
    I have a hash in which I store the products a customer buys (%orders). It uses the product code as key and has a reference to an array with the other info as value. At the end of the program, I have to rewrite the inventory to the updated version (i.e. subtract the quantity of the bought items) This is how I do rewrite the inventory: sub rewriteInventory{ open(FILE,'>inv.txt'); foreach $key(%inventory){ print FILE "$key\|$inventory{$key}[0]\|$inventory{$key}[1]\|$inventory{$key}[2]\n" } close(FILE); } where $inventory{$key}[x] is 0 → Title, 1 → price, 2 → quantity. The problem here is that when I look at inv.txt afterwards, I see things like this: CD-911|Lady Gaga - The Fame|15.99|21 ARRAY(0x145030c)||| BOOK-1453|The Da Vinci Code - Dan Brown|14.75|12 ARRAY(0x145bee4)||| Where do these ARRAY(0x145030c)||| entries come from? Or more important, how do I get rid of them?

    Read the article

  • My webservice works with soapclient.com but not with soapsonar software:

    - by Rebol Tutorial
    I have put a webservice I found here http://www.rebolforces.com/zine/rzine-1-02/#sect6. on my own website. I tested http://reboltutorial.com/discordian.wsdl with http://www.soapclient.com/soaptest.html it did work as I got this answer Sweetmorn, Discord 48, Year of Our Lady of Discord 3176 But doing the same thing with soapsonar gives me this response instead: <!DOCTYPE HTML PUBLIC "-//IETF//DTD HTML 2.0//EN"> <html><head> <title>301 Moved Permanently</title> </head><body> <h1>Moved Permanently</h1> <p>The document has moved <a href="http://reboltutorial.com/cgi-bin/discordian.cgi">here</a>.</p> </body></html> Is soapsonar buggy or did I mischief something ?

    Read the article

  • How do I mix functions in complex SSRS expressions?

    - by Boydski
    I'm writing a report against a data repository that has null values within some of the columns. The problem is building expressions is as temperamental as a hormonal old lady and doesn't like my mixing of functions. Here's an expression I've written that does not work if the data in the field is null/nothing: =IIF( IsNumeric(Fields!ADataField.Value), RunningValue( IIF( DatePart("q", Fields!CreatedOn.Value) = "2", Fields!ADataField.Value, 0 ), Sum, Nothing ), Sum(0) ) (Pseudocode) "If the data is valid and if the data was created in the second quarter of the year, add it to the overall Sum, otherwise, add zero to the sum." Looks pretty straight forward. And the individual pieces of the expression work by themselves. IE: IsNumeric(), DatePart(), etc. But when I put them all together, the expression throws an error. I've attempted about every permutation of what's shown above, all to no avail. Null values in Fields!ADataField.Value cause errors. Thoughts?

    Read the article

  • How to send an array(multidimensional/associative) and some varaibles together through json?

    - by I Like PHP
    i have an multidimensional array $ouptput=Array ( [0] => Array ( [0] => mov_1 [1] => MY FAIR LADY ) [1] => Array ( [1] => mov_10 [2] => love actually ) ) and two variables $avlblQnty=50 $success= true when i send these data via json echo json_encode( array('movieData'=>$output,'stock'=>$avlblQnty,'sucess'=>$success)); it returns {"movieData":[["mov_1","MY FAIR LAD],{"1":"mov_10","2":"love actually"}],"stock":0,"success":true} but i need json encoded data properly so that i can create an select box on movieData using(movieData.length), so for that i want json edcoded data in below format so that i can retrive successfully {"movieData":[{"mov_1":"MY FAIR LAD,mov_10":"love actually"}],"stock":0,"success":true} i want to know how to send an array(multidimensional/associative) and some varaibles together through json?

    Read the article

  • How to run xunit in Visual Studio 2012?

    - by user1978421
    I am very new to unit testing. I have been following the procedures for creating a unit test in visual studio 2012 on http://channel9.msdn.com/Events/TechEd/Europe/2012/DEV214. The test just won't start. And it will prompt me "A project with an Output Type of Class Library cannot be started directly. In order to debug this project, add an executable project to this solution which references the library project. Set an executable project as the startup project. Even though I attached the unit test class code to a console program, the test does not start and the test explorer is empty. In the video, it doesn't need to have any running program. The lady only created a class library, and the test will run. what should I do? Note. there is no "create unit test" on the mouse right click menu

    Read the article

  • disk-to-disk backup without costly backup redundancy?

    - by AaronLS
    A good backup strategy involves a combination of 1) disconnected backups/snapshots that will not be affected by bugs, viruses, and/or security breaches 2) geographically distributed backups to protect against local disasters 3) testing backups to ensure that they can be restored as needed Generally I take an onsite backup daily, and an offsite backup weekly, and do test restores periodically. In the rare circumstance that I need to restore files, I do some from the local backup. Should a catastrophic event destroy the servers and local backups, then the offsite weekly tape backup would be used to restore the files. I don't need multiple offsite backups with redundancy. I ALREADY HAVE REDUNDANCY THROUGH THE USE OF BOTH LOCAL AND REMOTE BACKUPS. I have recovery blocks and par files with the backups, so I already have protection against a small percentage of corrupt bits. I perform test restores to ensure the backups function properly. Should the remote backups experience a dataloss, I can replace them with one of the local backups. There are historical offsite backups as well, so if a dataloss was not noticed for a few weeks(such as a bug/security breach/virus), the data could be restored from an older backup. By doing this, the only scenario that poses a risk to complete data loss would be one where both the local, remote, and servers all experienced a data loss in the same time period. I'm willing to risk that happening since the odds of that trifecta negligibly small, and the data isn't THAT valuable to me. So I hope I have emphasized that I don't need redundancy in my offsite backups because I have covered all the bases. I know this exact technique is employed by numerous businesses. Of course there are some that take multiple offsite backups, because the data is so incredibly valuable that they don't even want to risk that trifecta disaster, but in the majority of cases the trifecta disaster is an accepted risk. I HAD TO COVER ALL THIS BECAUSE SOME PEOPLE DON'T READ!!! I think I have justified my backup strategy and the majority of businesses who use offsite tape backups do not have any additional redundancy beyond what is mentioned above(recovery blocks, par files, historical snapshots). Now I would like to eliminate the use of tapes for offsite backups, and instead use a backup service. Most however are extremely costly for $/gb/month storage. I don't mind paying for transfer bandwidth, but the cost of storage is way to high. All of them advertise that they maintain backups of the data, and I imagine they use RAID as well. Obviously if you were using them to host servers this would all be necessary, but for my scenario, I am simply replacing my offsite backups with such a service. So there is no need for RAID, and absolutely no value in another layer of backups of backups. My one and only question: "Are there online data-storage/backup services that do not use redundancy or offer backups(backups of my backups) as part of their packages, and thus are more reasonably priced?" NOT my question: "Is this a flawed strategy?" I don't care if you think this is a good strategy or not. I know it pretty standard. Very few people make an extra copy of their offsite backups. They already have local backups that they can use to replace the remote backups if something catastrophic happens at the remote site. Please limit your responses to the question posed. Sorry if I seem a little abrasive, but I had some trolls in my last post who didn't read my requirements nor my question, and were trying to go off answering a totally different question. I made it pretty clear, but didn't try to justify my strategy, because I didn't ask about whether my strategy was justifyable. So I apologize if this was lengthy, as it really didn't need to be, but since there are so many trolls here who try to sidetrack questions by responding without addressing the question at hand.

    Read the article

< Previous Page | 7 8 9 10 11 12 13 14 15 16 17 18  | Next Page >