Search Results

Search found 11618 results on 465 pages for 'shared storage'.

Page 138/465 | < Previous Page | 134 135 136 137 138 139 140 141 142 143 144 145  | Next Page >

  • Upgrading a dual-boot system HDD

    - by Jason
    I dual-boot my laptop due to lousy VM performance, and have a new 500GB/7200rpm drive coming in to replace the stock 320GB/5400rpm drive. I have the drive set up in three partitions: one for the Win7 system files, one for storage, and the third as the ext4 Linux file system. The system file and storage partitions are both NTFS. What I'm planning to do is use the system image creator built in Win7, then move that over to the new drive. However, how can I migrate the Ubuntu partition, and how do I make sure that the Grub bootloader isn't overwritten by the Windows loader?

    Read the article

  • TechEd 2014 Day 2

    - by John Paul Cook
    Today people asked me about backing up older versions of SQL Server to Azure. Older versions back to SQL Server 2005 can be easily backed up to Azure Storage by installing Microsoft SQL Server Backup to Windows Azure Tool. It installs a service of the same name that applies rules to SQL Server backups. You can tell the tool to backup or encrypt your SQL Server backups. You can have it automatically upload your backups to Azure Storage. Even if you don’t want to upload your backups to Azure, you might...(read more)

    Read the article

  • Oracle Solaris Cluster 4.1 Released

    - by Larry Wake
    Today we announced the release of Oracle Solaris Cluster 4.1 ( download ; existing customers can just update from the package repository ).New capabilities include:  Oracle Solaris 10 Zone Clusters: The easiest way to update and consolidate existing Solaris 10 application environments is with Oracle Solaris 10 Zones within Oracle Solaris 11 -- not only do you get higher system utilization, but you can immediately leverage new features such as network virtualization.With Oracle Solaris Cluster 4.1, you can now cluster these zones, for even higher availability. Expanded disaster recovery operations: Oracle Solaris Cluster 4.1 introduces managed switchover and disaster-recovery takeover of applications and data using ZFS Storage Appliance replication services in a multi-site, multi-cluster configuration. Faster application recovery with improved storage failure detection and resource dependency management. Labeled security support for providing both high availability and high security, leveraging Oracle Solaris 11 Trusted Extensions. Learn more: Oracle Solaris Cluster at the Oracle Technology Network Data Sheet  What's New in Oracle Solaris Cluster 4.1  FAQs

    Read the article

  • LAN or workgroup with Ubuntu server and windows clients

    - by Kenneth Fernando
    I have 35 windows7 standalone computers right now and planning to setup a LAN/workgroup using ubuntu as a server if its a LAN. Purpose for network.. Files from client computers to be accessed by me and vice versa through a shared folder on the server. Files only include word documents and very small project files. The network wont have Internet at the moment. What I would like to know.. How would i configure the ubuntu server to recognize the clients and will the clients be able to view the shared folders through their windows machines simultaneously? Would appreciate feedback. Thanks

    Read the article

  • Cloud Migration Lifecycle

    - by llaszews
    The first step in migrating to the cloud is the set up phase. This phase includes: 1. Plan Cloud Setup - Create a project plan. 2. Set up infrastructure - The IaaS and PaaS environments need to be installed and configured. This includes databases, network, hardware, disk and application servers. 3. Set up shared components - Are applications, database objects or other components going to be made accessible in the cloud ? These shared components need to be configured and made accessible. 4. Set up self-service portal - Customer self service provisioning, management and monitoring.

    Read the article

  • google-chrome without loading libnss3? [closed]

    - by 17763
    Possible Duplicate: How do I install Google Chome on legacy 7.10 computer? I am trying to get google-chrome to work on Ubuntu 7.10. I installed it with --force-depends and got it to install, but now when I try to run it, I get this error: /usr/bin/google-chrome: error while loading shared libraries: libnss3.so: cannot open shared object file: No such file or directory Is there a way to still get google-chrome to load even without this dependency satisfied? This is an old system that needs to keep this old 7.10 Ubuntu version and I would like to have google-chrome if possible installed, even if it means no sound or other features that are not compatible.

    Read the article

  • TechEd 2014 Day 2

    - by John Paul Cook
    Today people asked me about backing up older versions of SQL Server to Azure. Older versions back to SQL Server 2005 can be easily backed up to Azure Storage by installing Microsoft SQL Server Backup to Windows Azure Tool. It installs a service of the same name that applies rules to SQL Server backups. You can tell the tool to backup or encrypt your SQL Server backups. You can have it automatically upload your backups to Azure Storage. Even if you don’t want to upload your backups to Azure, you might...(read more)

    Read the article

  • Implementing RAID 1 in Ubuntu 10.04 Desktop [closed]

    - by Dibyendra
    I found many resources on implementing RAID 1 using two disk drives. But, I am confused while implementing RAID 1 using 4 RAID disks. Can we use two disks for storage and two for mirroring using RAID 1? I couldn't find the way to create RAID disk using gparted tool in Ubuntu 10.04 Desktop version. Maybe, the desktop version doesn't support RAID. I am trying to implement RAID on the existing Ubuntu installation? I have added 4 X 2TB HDD in the system and I want RAID 1 to be implemented in these 4 drives with 2 drives for storage and 2 devices for mirroring. Any help would be appreciated. Thanks! Updated: I installed Ubuntu 12.04 LTS and followed the following tutorial and it works now: www.youtube.com/watch?v=z84oBqOxsD0

    Read the article

  • Vertex data split into separate buffers or one one structure?

    - by kiba2
    Is it better to have all vertex data in one structure like this: class MyVertex { int x,y,z; int u,v; int normalx, normaly, normalz; } Or to have each component (location, normal, texture coordinates) in separate arrays/buffers? To me it always seemed logical to keep the data grouped together in one structure because they'd always be the same for each instance of a shared vertex and that seems to be true for things like character models (ex: the normal should be an average of adjacent normals for smooth lighting). One instance where this doesn't seem to work is other kinds of meshes like say a cube where the texture coordinates for each may be the same but that causes them to be different where the vertices are shared. Does everybody normally keep them separate? Won't this make them less space efficient if there needs to be an instance of texture coordinates and normals for each triangle vertex (They won't be indexed)? Can OpenGL even handle this mixing of indexed (for location) vs non-indexed buffers in the same VBO?

    Read the article

  • Samba user does not have folder read permission

    - by user289455
    I have set up a special user for read only samba shares. I set him up in Samba and as a system user. I shared a couple of folders but that user cannot access them. I know samba is working because I also shared them with the main user of the system which is an admin account and it works fine. How can I allow this user to have read permissions on all the directories I want to share without changing anything for any other users of the system? For example, I don't want to give him ownership of any of the files/directories. Just ongoing recursive read access. ongoing recursive is important. If someone adds a file or directory, I still want him to automatically be able to read it.

    Read the article

  • Core i7 with 4GB - go 64 bit or stay 32bit..?

    - by tommed
    I have a Core i7 laptop with 4gb ram. In windows, the disadvantage of using the 32bit OS would be that a single app could not use more than 2gb of ram (+ the 2gb shared kernel-space memory) except when setting a boot switch which reduces the amount of shared kernel memory, then it's 3GB max per app. What disadvantages/limitations would I have in Ubuntu for sticking to the 32bit OS? (If any?) - As I only have 4GB of ram, I can't see why I need to use the 64bit version?

    Read the article

  • How can cloudfare obfuscate IPs? [on hold]

    - by Sharen Eayrs
    I have several domains. I do not want those domains to be linked together. For example, say I have domaina.com. I do not want other users to be able to see that domaina.com is hosted with domainb.com A way to do it is to use many shared hosting service. Sometimes shared hosting is not powerful enough. Another way is to use multiple VPS, which is acceptable. My VPS provider says that I can accomplish this with cloudfare. However, I am confused on how cloudfare can hide my IP

    Read the article

  • How do I recover files from U1 cloud

    - by MarkWW
    OK I have just stopped syncing a folder. This appears to have been a huge mistake. I did not realise the folder would disappear from my cloud storage. I assume (and hope) the sub-folders/files it contained still exists in the cloud. I hope they still exists in the cloud because they do not exist on my computer. I did not delete these files from the cloud (I have tried to recover deleted files - none are being recovered). Where are they and how do I get them back into My storage?

    Read the article

  • Visual Studio Load Tests Virtual Users Simulation

    - by Eldar
    Hello, I'm currently working on writing a load testing application that takes advantage of Load Test using Visual Studio 2010. The load test will simulate 20 users on the same machine, and I need some data to be shared in-memory between all simulated users. I was suprised I couldn't find documentation answering the following question: What seperates each virtual user's running context from the other? Does each virtual user runs the tests in its own process? Maybe in its own app domain? Or just on its own thread? I need to know because if each user is running tests in its own process then all the in-memory cache isn't shared and is created for each user instead of one time for all of them, which is bad for me.

    Read the article

  • How to implement this .NET client and PHP server synchronization scenario?

    - by pbean
    I have a PHP webserver and a .NET mobile application. The .NET application needs data from a database, which is provided (for now) by the php webserver. I'm fairly new to this kind of scenario so I'm not sure what the best practices are. I ran into a couple of problems and I am not certain how to overcome them. For now, I have the following setup. I run a PHP SOAP server which has a couple of operations which simply retrieve data from the database. For this web service I have created a WSDL file. In Visual Studio I added a web reference to my project using the WSDL file and it generated some classes for it. I simply call something like "MyWebService.GetItems();" and I get an array of items in my .NET application, which come straight from the database. Next I also serialize all these retrieved objects to local (permanent) storage. I face a couple of challenges which I don't know how to resolve properly. The idea is for the mobile client to synchronize the data once (at the start of the day), before working, and then use the local storage throughout the day, and synchronize it back at the end of the day. At the moment all data is downloaded through SOAP, and not a subset (only what is needed). How would I know which new information should be sent to the client? Only the server knows what is new, but only the client knows for sure which data it already has. I seem to be doing double work now. The data which is transferred with SOAP basically already are serialized objects. But at the moment I first retrieve all objects through SOAP and the .NET framework automatically deserializes it. Then I serialize all data again myself. It would be more efficient to simply download it to storage, and then deserialize it. The mobile device does not have a lot of memory. In my test data this is not really a problem, but with potentially tenths of thousands of records I can imagine this will become a problem. Currently when I call the SOAP method, it loads all data into memory (as an array) but as described above perhaps it would be better to have it stored into storage directly, and only retrieve from storage the objects that are needed. But how would I store this? An array would serialize to one big XML (or binary) file from which I cannot choose which objects to load. Would I make (possible tenths of thousands) separate files? Also at the end of the day when I want to send the changes back to the server, how would I know which objects to send... since they're not in memory? I hope my troubles are clear and I hope you are able to help me figure out how to implement this the best way. :) Some things are already fixed (like using .NET on the mobile device, using PHP and having a MySQL database server) but other things can most certainly be changed (like using SOAP).

    Read the article

  • What is the good way of sharing specific data between ViewModels

    - by voroninp
    We have IAppContext which is injected into ViewModel. This service contains shared data: global filters and other application wide properties. But there are cases when data is very specific. For example one VM implements Master and the second one - Details of selected tree item. Thus DetailsVm must know about the selected item and its changes. We can store this information either in IAppContext or inside each concerned VM. In both cases update notifications are sent via Messenger. I see pros and cons for any of the approaches and can not decide which one is better. 1st: + explicitly exposed shared proerties, easy to follow dependencies - IAppContxt becomes cluttered with very specific data. 2nd: the exact opposite of the first and more memory load due to data duplication. May be someone can offer design alternatives or tell that one of the variants is objectively superior to the other cause I miss something important?

    Read the article

  • Alternatives to using cookies?

    - by theclueless1
    Whate are alternatives to using cookies/client-side storage for a PHP/MySQL based site on Apache. Scenario/Requirements: I want to try using some anti-bot code to prevent specific scrapers etc. from accessing the site. I would like to run this code before launching the rest of the site (before DB access etc.). I don't want to constantly run the same code on every page-load after a visitor has passed the initial check. I'd like to avoid the use of Cookies/Client side storage if at all possible. The only solution I can currently think of is to write files to the server based on the visitors IP/UA, or to write a list of them to a single file. Yet this has the limitation of multiple users through a proxy/same connection, etc ... So, any ideas/suggestions? Or am I simply over working the issue?

    Read the article

  • YouSendIt Alternative?

    - by user4855
    Looking for a reasonably priced alternative to YouSendIt's exorbitant pricing for an embedded, unbranded (i.e. no "Uploads by SomeCompany" or at the very least, discrete, subtle co-branding) file upload solution for my client's print shop Website. To do what we want to do with YouSendIt, we're looking at a corporate account of $995 USD plus $29.99 USD monthly fee, that is only sold pro-rated, so you have to buy the entire year's worth. To me, this is just unacceptable considering the commodity pricing of storage and bandwidth nowadays. For data, we're looking at roughly 10MB per upload, with perhaps 250-1000 uploads per month, with transient data storage of no more than 30 days (and more than likely 1-2 business days) for a total of 10 GB transfer (upload) and 10 GB transfer (download, to the print shop) at the very max each month. Any ideas? Everything I've found through searching seems to be geared more towards personal file sharing and not for embedding into Websites. Thanks

    Read the article

< Previous Page | 134 135 136 137 138 139 140 141 142 143 144 145  | Next Page >