Search Results

Search found 9847 results on 394 pages for 'cloud backup'.

Page 53/394 | < Previous Page | 49 50 51 52 53 54 55 56 57 58 59 60  | Next Page >

  • WebLogic & Coherence & Cloud presentations for customer meetings

    - by JuergenKress
    We updated the WebLogic Community Workspace with the latest customer facing presentations: WebLogic 12c.pptx Oracle Coherence.pptx CAF WebLogic Server 12c, Coherence and Java EE.pptx Cloud Your Strategy Your Cloud Your Choice.pptx For all presentation, please visit the WebLogic Community Workspace (WebLogic Community membership required). WebLogic Partner Community For regular information become a member in the WebLogic Partner Community please visit: http://www.oracle.com/partners/goto/wls-emea ( OPN account required). If you need support with your account please contact the Oracle Partner Business Center. Blog Twitter LinkedIn Mix Forum Wiki Technorati Tags: WebLogic,Coherence,Cloud,presentations,ppt,sales,WebLogic presentation,WebLogic Community,Oracle,OPN,Jürgen Kress

    Read the article

  • Google I/O 2012 - Gaming in the Cloud

    Google I/O 2012 - Gaming in the Cloud "Fred Sauer Many games developers are finding the easy development and deployment experience of Google App Engine ideal for building cloud based state-storage, matching making services and collaborations services. When you have a hit game, the last thing you want to do is worry about your server provisioning. App Engine has an always-free tier to get you started and then scales seamlessly to any size of usage. Game developers also use Google Cloud Storage to easily store and quickly deliver media files to clients around the world. For all I/O 2012 sessions, go to developers.google.com From: GoogleDevelopers Views: 1 0 ratings Time: 01:02:17 More in Science & Technology

    Read the article

  • FMW Cloud Forum: Chicago

    - by kellsey.ruppel
      The increasing popularity of cloud computing is changing how enterprise systems are managed and organized--and that change does not stop at the datacenter. Cloud computing is also changing how enterprises develop and build business applications, a shift that will require unprecedented collaboration across the enterprise, from developers to the user community. Are you currently building applications in the Cloud? What concerns or challenges do you forsee in doing so? Oracle experts will be discussing these topics and how with a user experience platform you can leverage new collaborative practices to design and build applications that deliver business value and meet exacting user requirements. Join us in Chicago on June 29th to learn more and hear from Oracle experts. Not located in Chicago? We're coming to a city near you!

    Read the article

  • Forbes.com: Oracle's message is Loud & Clear – “We’ve Got The Cloud”

    - by Cinzia Mascanzoni
    In a two-part series on Oracle's cloud strategy, Bob Evans reports on the October 4 meeting where Wall Street analysts questioned Mark Hurd and Safra Catz about the company's positioning for the shift to cloud computing. Access the article and read the Q&A exchanges between the analysts and Hurd and Catz. And then check out Bob's related Forbes.com piece "The Dumbest Idea of 2013," in response to the preposterous chatter that Larry Ellison and Oracle don't "get" the cloud. His powerful six-point argument unravels our competitors' spin. Go to the two-part strategy article. Read the "Dumbest Idea." Follow Bob on Twitter as he frequently updates his Oracle Voice column on Forbes.com.

    Read the article

  • Backup and Recovery in Exadata environments

    - by Javier Puerta
    As with any infrastructure every Engineered Systems customer needs a Backup & Recovery solution for Data Protection. See a detailed presentation and learn about the challenges of backup & recovery and the key benefits of the ZFS Storage Applicance as a backup device for Exadata & Sparc SuperCluster. (You need to be a registered member of the Exadata Partner Community to access link above. Otherwise you will get an error. You can register here)

    Read the article

  • Log backups "stalling" on SQL 2008?

    - by MattK
    I have interited a box running SQL Server 2008 and Windows 2003, and have had a few events where largeish (35GB) log backups "stall", both before and after the installation of SQL 2008 SP1. The server log ships to a standby, so regular log backups are taken at 15 minute intervals. However, after an index reorg causes the log to grow to about 35GB (on a DB with about 17GB of data), the next log backup runs to ~95% completion, then seems to stop. The process shows as suspended, with a wait state of BACKUPIO. CPU, read, and write activity on the SPID also does not change, and the process stays in this state for hours, when normally a backup of this size should complete in about 20 minutes. This server has a single RAID-1 volume, thus the source database files and destination backup files are on the same volume. However, I cannot determine if another process is blocking the backup. The backup SPID cannot be killed, and the only way to terminate the log backup and clear the lock on the backup file is to cycle the SQL Server service. There was one event where the backup terminated completely, with an error that another process had locked the backup file, but no details about what that process was. Can anyone suggest a cause or diagnostic process to this situation?

    Read the article

  • TFS Backup Plan Wizard Tool

    - by Enrique Lima
    With the release of the “September – 2010” TFS 2010 Power Tools, came an addition to the Team Foundation Server Administration Console.  This addition is the Team Foundation Backups Tree item.  The tool is used to create backup plans and to work with it you run through a wizard, just like you would in configuring TFS or any of the extensions it has. The areas covered through the tool include: Backup to a Network Backup Path, retention configuration. Under Advanced Options, the extension to be used for the Full and Transactional backups. The capability to include external databases, meaning, include the reporting databases and SharePoint databases as part of the plan. There are further options as you can see, that includes being able to define a task scheduler account, be able to set alerts for notifications on execution of the plans, and last the option to configure the schedule for the plan execution.  All in all a very good tool and great way to safeguard the investment you’ve made.

    Read the article

  • SQL Server log backups “stalling”

    - by MattK
    I have interited a box running SQL Server 2008 and Windows 2003, and have had a few events where largeish (35GB) log backups "stall", both before and after the installation of SQL 2008 SP1. The server log ships to a standby, so regular log backups are taken at 15 minute intervals. However, after an index reorg causes the log to grow to about 35GB (on a DB with about 17GB of data), the next log backup runs to ~95% completion, then seems to stop. The process shows as suspended, with a wait state of BACKUPIO. CPU, read, and write activity on the SPID also does not change, and the process stays in this state for hours, when normally a backup of this size should complete in about 20 minutes. This server has a single RAID-1 volume, thus the source database files and destination backup files are on the same volume. However, I cannot determine if another process is blocking the backup. The backup SPID cannot be killed, and the only way to terminate the log backup and clear the lock on the backup file is to cycle the SQL Server service. There was one event where the backup terminated completely, with an error that another process had locked the backup file, but no details about what that process was. Can anyone suggest a cause or diagnostic process to this situation?

    Read the article

  • Cloud Migration Lifecycle

    - by llaszews
    The first step in migrating to the cloud is the set up phase. This phase includes: 1. Plan Cloud Setup - Create a project plan. 2. Set up infrastructure - The IaaS and PaaS environments need to be installed and configured. This includes databases, network, hardware, disk and application servers. 3. Set up shared components - Are applications, database objects or other components going to be made accessible in the cloud ? These shared components need to be configured and made accessible. 4. Set up self-service portal - Customer self service provisioning, management and monitoring.

    Read the article

  • Cloud consolidation handling multi databases

    - by llaszews
    I have spoken about virtualization and the different types of virtualization. Which includes OS zones, application server domains, database schemas, VLANS and other approaches. Another approach is to create a virtually federated database in the cloud. DBSpaces is a company that has a technology to created a virtually federated database in the cloud. DBSpaces is a Virtual Database technology that allows an organisation thru a single Virtual Database access multiple data sources (or database spaces) in real-time. Additionally dbSpaces can be configured to access an organisations data internally using a remote gateway so that their dbSpace is seamless across the Public and Private cloud.

    Read the article

  • Oracle RightNow Cloud Service Roadmap - Live Webcast, Nov 13, 6pm CET

    - by Richard Lefebvre
    Did you miss out on Oracle OpenWorld this year? Then make sure you don’t miss out on this Webinar.  The Oracle RightNow development team shares the latest innovations and integrations, including future roadmap, for the Oracle RightNow Service Cloud platform.  Find out how these innovations will help you deliver exceptional customer service to your customers.Join our live Webcast on Wednesday, November 13, 2013, 09:00 a.m. PT / 12:00 p.m. ET (18:00 p.m. CET) to get up to speed with the latest updates and future capabilities, and learn how you can: Provide a more engaging Web experience Increase the effectiveness of assisted service Deliver the right answer at the right time - all the time Agenda Topics How RightNow fits into Oracle's vision for Oracle Service The latest updates to the Oracle RightNow Cloud Service platform Oracle's key investment areas and roadmap for Oracle RightNow Cloud Service Don’t miss this chance to learn how you can delight your customers while improving cost and efficiency.Register Now

    Read the article

  • Backing up 80G hard drive 1G per day

    - by barrycarter
    I want to securely backup my 80G HD, but doing a complete backup takes forever and slows down my machine, so I want to backup just 1G per day. Details: % First hurdle: on the first day, I want to backup the "first" 1G of the hard drive. Of course, there really is no "first" 1G on a hard drive. % After 80 days, I'll have my whole HD backed up... assuming none of my files ever change, which of course they do. So the backup plan/program must also catch file creation/changes as they come along. % The backups must be consistent, in that I can restore my system by restoring the backups sequentially. In other words, "dd if=/harddrive" probably won't work. % The backups should encrypt file contents AND names, but I don't see this as a major hurdle. % Once the backup has backed up everything (even changed files), it can re-backup the first 1G on my hard drive. Even though this backup is redundant, that's OK, because I always want to be backing up something (eg, if I'm backing up to optical media, the older media might start going corrupt). Is there a magic backup plan/program that does this? In reality, I want to do this for multiple machines with multiple drives each, but think that solving the above will solve the general case.

    Read the article

  • Quels sont les avantages du Cloud Computing pour votre entreprise ? Google lance "Go Google Cloud Ca

    Quels sont les avantages du Cloud Computing pour votre entreprise ? Google lance "Go Google Cloud Calculator" pour en appréhender les bénéfices Le cloud computing intéresse de plus en plus d'entreprises, et des millions d'entre elles se sont tournées vers Google en migrant vers les Google Apps. Une décision encore audacieuse, tant il est difficile d'imaginer ce que le travail « dans le nuage » signifie vraiment et quels en sont les véritables avantages. Quel est l'impact de la collaboration en ligne sur votre lieu de travail ? Comment une capacité de stockage d'email accrue ou une messagerie instantanée intégrée et le chat vidéo pourraient avoir un impact sur la productivité de votre entreprise ?

    Read the article

  • Why Deliver Customer Service in the Cloud?

    - by Charles Knapp
    In volatile, competitive markets, delivering exceptional service across channels is essential. But delivering world-class service on tight budgets, and deliving improvements quickly, is a tough challenge. That's why so many of the world's most successful organizations choose to deliver customer service in the cloud. Example: Michele Watson, VP of Global Customer Care at Match.com, says Oracle's service in the cloud "helps our customer receive the support they need in real time, our contact center agents be more productive and helpful, and our executive and product development teams receive detailed feedback to continue our improve our customers' experience." Learn more here about why you should consider delivering customer service in the cloud

    Read the article

  • How Can I Automate the Backup of a Quickbooks Server?

    - by Nick
    I have three computers: The first is the company file server which has the Quick Books company file, is always on, and lives in the closet. The other two are Quick Books Clients. All are XP Pro. I need a way to automatically backup the QB data file, without any user intervention. Quick Books has a built in scheduled backup utility, but from what I've read, it only works when the software is running in single user mode. (and obviously putting the server into single user mode defeats the concept of having a server). Also, I'm not actually running QB itself on the server, just the "QB Database Server" process that sits in the system tray. Surely there must be a way to automate this? I'm open to any ideas/suggestions. Thanks!

    Read the article

  • How to backup/restore excluding filestream varbinary in SQL Server 2008?

    - by fdierre
    There is an application used in a production site that uses SQL Server 2008 as its DBMS. The database schema uses Filestream Varbinary to save binary data on the filesystem instead of directly into the DB tables. The point is that now and then it would be useful to copy the production database on development machines, mostly for doing troubleshooting. The database is too big for comfortably moving it around, but it would be ok if it could be moved leaving out the filestream varbinary fields. In other words, I am trying to make an "imperfect" copy of a database: i.e., on the destination database, it is ok to have NULL values instead of the varbinary. Is this possible? I tried looking for the feature on the SQL Server Management studio and did a backup that excludes the filegroup containing the filestream varbinary, but I cannot restore: MSSMS complains that the restore cannot be done because the backup is incomplete (of course). Is it possible to achieve what I am trying to do in some way?

    Read the article

  • "tar -cfz" versus "tar cf - | gzip": are they different? (or how to improve a backup)

    - by I'm Dario
    I want to speed up my backup done with "tar -cfz", the common way to do it. But day by day my backed up files grow so it becomes slower. I was thinking to take advantage of the several cores available in my server and I was wondering if there is any difference between doing the backup with "tar -cfz" or piping tar to gzip ("tar cf - | gzip"). I guess that there isn't any difference, because the first spawns two processes (tar and gzip), in a similar way like piping it. If there is not difference, do you know any good alternative to do this, without going incremental? I'm looking at pigz too and it looks fine.

    Read the article

  • Client backup solution for small (100-150 user) homogenous win/nix/mac office?

    - by Gomibushi
    We are currently using Symantec Backup Exec with Desktop and Laptop Option for our Windows clients, time machine for mac and offer simple rsync to linux users, in addition to home folders that are always backed up and available. We are not overly happy with the horrid complexity and multitude of minor bugs in SBE, but "when you don't touch it, it mostly works". Ideally we'd like to offer a real and full backup solution to all clients, but mostly to Linux users, as they don't have a good alternative. I have barely tested Druva on windows, and it is promising in its simplicity and "it just works" looks, but does anyone have experience with it? This post lists some that I will look at.

    Read the article

  • Is there a simple way to backup and restore all Microsoft SQL Server database objects related to a p

    - by Nathan Hartley
    I would like to backup, not only the databases that belong to a particular application living on a shared server, but also, those things that get stored outside of the database; the server accounts, jobs, maintenance plans and whatever else I can't think of at the moment. This backup should be complete enough that it's corresponding restore will recreate the entire application on a different SQL server. This seems like a problem others must have dealt with in the past. So before I embark on creating custom Powershell scripts for each application, I have come to ask you... Can you help?

    Read the article

  • How do I do a cross-platform backup/restore of a DB2 database?

    - by Pridkett
    I need to dump a couple of databases from DB2 for Mac and DB2 for Linux and then import the databases to DB2 for Windows. Unfortunately, when I try the standard backup and restore I get the following error: SQL2570N An attempt to restore on target OS "NT-32" from a backup created on source OS "?" failed due to the incompatability of operating systems or an incorrect specification of the restore command. Reason-code: "1". I've seen references to DB2 needing an IXF dump and import, but I can't find any solid information about how to do this without dozens of other steps. Any hints on how to do this in the least painful manner?

    Read the article

  • Best practice? Using DPM to backup VMs within each VM or through the host?

    - by andrew
    We've got two Hyper-V hosts running multiple VMs (all flavors of Windows Servers). One of the VMs is running MS Data Protection Manager 2010, which runs beautifully (most of the time) and is connected to a separate NAS via iSCSI for the DPM storage. I noticed when I installed the DPM agent on the Hyper-V hosts, it enumerates the VMs in the DPM Protection listing. I don't want to burn through my storage space too fast with duplicate protection, so I was wondering: Is it recommended to back up VMs through the host, or is it better to install the DPM agent on each VM and backup as I would any other machine? It would seem as though most people (currently including me) do it the second way, but is there any advantage to including the entries under HyperV (Backup using Child Partition Snapshop)?

    Read the article

  • How do I backup my Ubuntu 9.10 system and then restore it to a new machine with different hardware?

    - by EricJLN
    I have a nicely configured Ubuntu 9.10 machine, with crossover linux (from Codeweavers) installed. I have the Nvidia drivers installed. Everything is just as I want it. Now I want to move to a new machine with different hardware: different sized hard drive, different size RAM, different video card, etc. I tried the technique suggested on Ubuntu forums, "Howto: Backup and restore your system!". The results: New system complains about swap not being found New system complains about not finding Nvidia hardware, and I can't open the System-Administration-Hardware Drivers application (i.e., /usr/bin/jockey-gtk) to fix the problem. As result of nvidia, X-Windows is throwing errors, and the on screen process for fixing X-Windows isn't working. How do I restore a backup of my old system, including my Crossover Linux installation, to a new machine with different hardware installed?

    Read the article

  • How to backup data on debian vps to dropbox?

    - by IBr
    I have really simple private VPS with some webpages and music server. I want to backup some configs and some scripts to dropbox or similar service. Server has no gui (except simple ssh X forwarding, which is neither convenient for constant usage and does not provide full desktop) everything is controlled through ssh. So my question would is it possible to setup dropbox client for command line use? How? Is there any alternatives for dropbox, which would have command line clients? Also is it possible to incorporate backup into script for cron job?

    Read the article

  • How to properly backup mediawiki database (mysql) without messing up the data?

    - by Toto
    I want to backup a mediawiki database stored in a MySQL server 5.1.36 using mysqldump. Most of the wiki articles are written in spanish and a don't want to mess up with it by creating the dump with the wrong character set. mysql> status -------------- ... Current database: wikidb Current user: root@localhost ... Server version: 5.1.36-community-log MySQL Community Server (GPL) .... Server characterset: latin1 Db characterset: utf8 Client characterset: latin1 Conn. characterset: latin1 ... Using the following command: mysql> show create table text; I see that the table create statement set the charset to binary: CREATE TABLE `text` ( `old_id` int(10) unsigned NOT NULL AUTO_INCREMENT, `old_text` mediumblob NOT NULL, `old_flags` tinyblob NOT NULL, PRIMARY KEY (`old_id`) ) ENGINE=InnoDB AUTO_INCREMENT=317 DEFAULT CHARSET=binary MAX_ROWS=10000000 AVG_ROW_LENGTH=10240 How should I use mysqldump to properly generate a backup for that database?

    Read the article

  • backup sql databases, folders; 7ZIP and copy to ftp

    - by laurens
    Hi all, We are quite stuck with which solution to choose for this backup issue: What should happen: First, there should be an interface were to choose several sql databases (sort of checkboxes or whatsoever), also a few folders should be backed up - this could be part of the program or could be seperate, I think about an interface were to select folders, but a txt file (or xml) with paths-to-folders is as good. Next, everything should be 7Zipped, SQL-DB and files seperate. Eventually everything should be copied to a local network drive after which copied via FTP. Also important; it could be programmed or (partly) bought but I can't be one of those expensive backup tools $1000+ etc. I already found this fairly priced tool that does already most of the tasks 7ZIP and copy to ftp sqlbackupandftp.com For your information: we had a kind of self-made tool created by a collegue (some time ago) but it became very untrustworthy and as the databases grew it couldn't handle it anymore... moving on Please come up with suggestions. Thanks in advance!

    Read the article

< Previous Page | 49 50 51 52 53 54 55 56 57 58 59 60  | Next Page >