Search Results

Search found 13575 results on 543 pages for 'desktop virtualization'.

Page 37/543 | < Previous Page | 33 34 35 36 37 38 39 40 41 42 43 44  | Next Page >

  • Use desktop widget to run corporate web application

    - by jonny
    So the question is simple: is it possible to set windows desktop to corporate application page (anybody remembers that active desktop thing?) and use it instead of browser (sic!). My thoughts: Windows 7: I can create giant full-screen widget that will show my page. Still have no idea if I will be able to navigate inside it though. Windows Vista: ???. I know there is still a side bar gadget that hosts a browser. Windows XP: create static html page with iframe pointing to corporate app home page. Problem: clicking links on page still opens default browser.

    Read the article

  • Creating a Linux Desktop Envoriment

    - by Alon
    Suppose I want to create my own desktop envoriment for Linux, without X. Like Google with the Android did. Where do I start? Is it actually a normal application that just draws stuff, and starts after the kernel boot? And how does it draw it? Using OpenGL or is there something more generic? And graphics drivers, how is it going? You should develop custom graphics drivers for your desktop or it comes with the Linux kernel? Note: It's for normal PCs and not embedded devices. Thanks.

    Read the article

  • Developing games using virtualization on macOS (or Linux) [on hold]

    - by zpinner
    From what I've seen, most of the gamedev tools and engines (that could generate cross platform games) are not supported on Mac. Havok/Project Anarchy, UDK, GameMaker, e.g. . Basically, the only options I found are: Unity3d and monogame + xamarin. Unity is nice and I've been playing with it for some time, but the free version is quite limited when we're talking about shaders, that made me consider that as an indie developer, I might want more freedom to experiment new things, without paying the expensive unity license. I didn't try monogame + xamarin yet, and altough XNA is a very nice game framework, I'd like to have more freedom to experiment and finish a game first before paying for the IDE, which is not possible with the current Xamarin business model. That leaves me with the thought that I must go back to windows, which I'd preferably do it partially, if it's possible. Using BootCamp is something that I'd like to avoid, since it's a pain to reboot when changing OS and that would probably force me to become a 100% windows user. Is there anyone actually developing a game using virtualization solutions like parallels or vmwareFusion? How was your experience?

    Read the article

  • Virtualization or Raw Metal?

    - by THE
    Normal 0 21 false false false DE X-NONE X-NONE MicrosoftInternetExplorer4 /* Style Definitions */ table.MsoNormalTable {mso-style-name:"Table Normal"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-qformat:yes; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:11.0pt; font-family:"Calibri","sans-serif"; mso-ascii-font-family:Calibri; mso-ascii-theme-font:minor-latin; mso-hansi-font-family:Calibri; mso-hansi-theme-font:minor-latin; mso-fareast-language:EN-US;} With the growing number of customers who want to run the Oracles EPM/BI (or other Fusion Middleware Software) in a virtualized environment, we face a growing number of people asking if running Oracle Software within VMware is supported or not. Two KM articles reflect Oracles policy towards the use of VMware: 249212.1 and 475484.1 . The bottom line is: “you may use it at your own risk, but Oracle does not recommend it”. So far we have seen few problems with the use of VMware (other than performance and the usual limitations) but Oracle does not certify its software for the use in VMware (and specifically for RAC Software actively refuses any support) and any issue that may occur will be fixed for the native OS only. It is on the customer to prove that the issue is NOT due to VMware in case that an issue is encountered. See: “Oracle Fusion Middleware Supported System Configurations page” And also “Supported Virtualization and Partitioning Technologies for Oracle Fusion Middleware”

    Read the article

  • split virtualization design based on environment or server role?

    - by Dan
    I'm setting up the server environment for a new software development group, which will include 4 test environments. These are web applications, so each environment will have an application server and a database server. I'm planning on buying two physical servers (e.g. 6-core CPU each with 12GB or so of RAM), and I'm thinking virtualization is appropriate here. With that in mind, I've thought of a couple ways that I could organize the virtualization strategy: - Separated by server role: Server 1 has all the application servers, each in their own guest VM. Server 2 has all the databases. OR - Separated by environment: Server 1 has a VM for two of the environments, with the VM containing both the app server and the database server. Server 2 would also contain two test environments, with the same style (app server and database in same VM). The advantages I see with all the app servers on one server and all the databases on another server is that I could probably be more efficient with the database server (one instance running multiple databases). But the other option seems easier to manage (archives/restorations would be contained in a single VM). Any recommendations? TIA.

    Read the article

  • How can I generate filesystem images that are usable on many different virtualization systems?

    - by Mark Longair
    I have written a script that generates a root filesystem image (based on Debian lenny) suitable for User-Mode Linux. (Essentially this script creates a filesystem image, mounts it with a loop device, uses debootstrap to create a lenny install, sets up a static IP for TUN/TAP networking, adds public keys for login by SSH and installs a web application.) These filesystem images work pretty well with UML, but it would be nice to be able to generate similar images that people can use on alternative virtualization software, and I'm not familiar with these options at all. In particular, since the idea is to use this image as a standalone server for testing the web application, it's important that the networking works. I wonder if anyone can suggest what would be involved in customizing such root filesystem images such that they could be used with other virtualization software, such as VMware, Xen or as an Amazon EC2 instance? Two particular concerns are: If such systems don't use a raw filesystem image (e.g. they need headers with metadata or are compressed in some particular way) do there exist tools to convert between the different formats? I assume that in the filesystem, at least /etc/network/interfaces will have to be customized, but are more involved changes likely to be necessary? Many thanks for any suggestions...

    Read the article

  • Windows 8 client virtualization

    - by John Paul Cook
    Hyper-V is coming to Windows 8, but you must have a processor that supports SLAT. Virtual machines created with Virtual PC aren’t easily transferred to Windows 2008 Hyper-V and vice-versa. With Windows 8, it will be easy to move vhds from Windows 8 on your laptop or desktop to Windows 8 server and back again. To find out if your processor supports SLAT, run coreinfo –v from a command window running as administrator. Download coreinfo from here . My MacBook Pro supports SLAT as this output shows:...(read more)

    Read the article

  • London Nov-8: Desktop Virtualisation Seminar

    - by mprove
    >> Simplify Application and Data Access with Oracle Desktop VirtualisationMany companies claim they’ll handle your application access needs, and yet only Oracle can provide you with every component needed for secure and reliable access to Oracle Applications and other enterprise software from a variety of devices. This means you can design your deployment knowing that all of the pieces work together, from applications and virtualisation to servers and storage systems.Join us to learn how Oracle desktop virtualisation helps you get the most from your valuable IT resources. Topics we’ll cover and demonstrate in this productive half-day event include: How to provide secure access to applications and data from nearly anywhere on a wide range of devices Use cases for desktop virtualisation How desktop virtualisation can support a wider business transformation agenda Reasons to embrace employees using their own devices for work-related activities How virtualisation can extend the life of your PCs and other devices How desktop virtualisation can decrease your carbon footprint and IT costs << Register here for the free event

    Read the article

  • Advantages of Using Linux as primary developer desktop

    - by Nick N
    I want to get some input on some of the advantages of why developers should and need to use Linux as their primary development desktop on a daily basic as opposed to using Windows. This is particulary helpful when your Dev, QA, and Production environments are Linux. The current analogy that I keep coming back to is. If I build my demo car as a Ford Escort, but my project car is a Ford Mustang, it doesn't make sense at all. I'm currently at an IT department that allows dual boot with Windows and Linux, but some run Linux while the vast majority use Windows. Here's several advantages that I've came up with since using Linux as a primary desktop. Same Exact operating system as Dev, QA, and Production Same Scripts (.sh) instead of maintaining (.bat and *.sh). Somewhat mitigated by using cygwin, but still a bit different. Team learns simple commands such as: cd, ls, cat, top Team learns Advanced commands like: pkill, pgrep, chmod, su, sudo, ssh, scp Full access to installs typically for Linux, such as RPM, DEB installs just like the target environments. The list could go on and on, but I want to get some feedback of anything that I may have missed, or even any disadvantages (of course there are some). To me it makes sense to migrate an entire team over to using Linux, and using Virtual Box, running Windows XP VM's to test functional items that 95% of most of the world uses. This is similar but a little different thread going on here as well. link text

    Read the article

  • Change workarea size of Linux desktop

    - by nonoitall
    I'm trying to write a taskbar/panel for Linux (like fbpanel or pypanel) using GTK# and am a little hung up. I've created a Gtk.Window to act as the panel and positioned/resized it appropriately. I've also set its WindowTypeHint to Dock so that it remains on top of other windows. So far it 'looks' like a panel. However, if the panel is running and I maximize another window, that window fills the whole desktop - meaning the bottom portion of the window is covered up by my panel. I've gathered that I probably need to change the desktop's workarea. How can I go about doing this in C#? (Preferably using GTK#, but I don't mind using interop if it's necessary.) As a bit of a side point, I'm curious if anyone knows how I would go about 'informing' the window manager about where applications' taskbar buttons are. (For example, if the window manager wants to animate the minimize action so that the window shrinks down to its button on the taskbar, how do I let the window manager know where that button is on the taskbar?)

    Read the article

  • What is the best strategy for licensing a desktop application using a web service, when all I need to know is when people use the product?

    - by user1667022
    Our company's main application is a desktop program that is used at warehouses and written in C# and Windows Presentation Forms. The next thing we want to be able to do is track when customers open up the application and when it is being used. The reason for this is so we can charge them per month, based on if they are/arn't using the application. My boss is having me research different ways to "license" the product under these requirements. Not having any experience doing this, a few things come to mind. I could create a web application that runs on a server, and every time the desktop application is opened and the user logs in, the application connects to the server and marks a database with the DateTime. Or is there licensing software that I can use to accomplish this? Just looking for tips/advice from people who have experience with this type of stuff.

    Read the article

  • Which virtualization solutions are available for non-x86 platforms?

    - by asmaier
    Are their any virtualization solutions like Xen, KVM and VMWare ESX available for non-x86 platform? Especially I'm interested if there are solutions available or if Xen, KVM can be made to run on the platforms POWER5/6/7 PowerPC Itanium 64 NEC SX-8/9 Cray X2 BlueGene What are your experiences? Is it possible to virtualize GPUs like Nvidias Tesla/Fermi?

    Read the article

  • How to configure Remote desktop on window server 2008 R2?

    - by Abdullah BaMusa
    I’m trying to connect over internet to my home workstation which has Windows Server 2008 R2 (Web Edition) installed from my PC at work (Windows 7 installed on it) via Remote Desktop. I configure the workstation to accept remote desktop and I can connect to it from my laptop if I’m within same Home LAN but I can’t establish the connection from my PC at work . My question is: Is possible to connect to my workstation over internet using remote desktop? Is there any step by step resource the setup this feature?

    Read the article

  • How to configure Remote desktop on window server 2008 R2?

    - by Abdullah BaMusa
    I’m trying to connect over internet to my home workstation which has Windows Server 2008 R2 (Web Edition) installed from my PC at work (Windows 7 installed on it) via Remote Desktop. I configure the workstation to accept remote desktop and I can connect to it from my laptop if I’m within same Home LAN but I can’t establish the connection from my PC at work . My question is: Is possible to connect to my workstation over internet using remote desktop? Is there any step by step resource the setup this feature?

    Read the article

  • How to configure Remote desktop on window server 2008 R2?

    - by Abdullah BaMusa
    I’m trying to connect over internet to my home workstation which has Windows Server 2008 R2 (Web Edition) installed from my PC at work (Windows 7 installed on it) via Remote Desktop. I configure the workstation to accept remote desktop and I can connect to it from my laptop if I’m within same Home LAN but I can’t establish the connection from my PC at work . My question is: Is possible to connect to my workstation over internet using remote desktop? Is there any step by step resource the setup this feature?

    Read the article

  • How to prevent any files from being copied onto the Desktop?

    - by Kaushik Gopal
    Gentlepeeps, Is there any solution to prevent copying of files onto the desktop? My folks and i currently have one super-common-comp in our place. And when all of the laptops are busy, we tend to do some quick surfing on this comp. I prefer to keep things freakishly organized in this comp and one thing i hate is having those off-shoot irritating forwarded downloads on my desktop. Is there some way to restrict files from being copied onto the desktop? I tried making the administrator user's desktop folder as read-only :P . That doesn't solve the problem. Any other ideas/software/solutions? cheers.

    Read the article

  • How can I have Windows 8 go to the desktop by default?

    - by Schnapple
    I've played around a smidge with the Windows 8 Consumer Preview in a VirtualBox VM and I think the improvements under the hood may be worth tolerating the Metro UI crap. I don't like that the entire screen changes to something I don't care about when I hit the Windows key and start typing but I can deal with it. The one thing I cannot stand, though, is that it starts in the Metro interface by default. I have to hit the "Desktop" tile to get to a normal interface, and while I can hit "escape" to go back to the desktop and dismiss Metro, it doesn't work when you first log in. When you first log in, you have to hit that "Desktop" tile. I know that the Enterprise versions of Windows 8 will go to the desktop by default. I don't know but I would assume that there's probably some registry key that would handle this. Has anyone figured that out yet?

    Read the article

  • What software will run the new Windows 8 in virtualization under OS X?

    - by bmike
    I've read a lot about the new Windows OS, but would like to run it first hand. I've got several Macs with Mountain Lion (10.8) and presume I can use this evaluation license to test the OS for my consulting company. What else do I need to run Windows 8 in virtualization (as I'd rather not mess with BootCamp for several reasons)? Any other gotchas from someone who has gone down this path would be appreciated as well.

    Read the article

  • Remote desktop logs in with saved credentials when connecting from one computer but asks for username and password from another.

    - by AndrejaKo
    I have two computers on which I recently installed windows 7 64 bit. Using VPN, they connect to remote network on which there is a Windows Server 2000 SP4 computer. User is supposed to connect to server using remote desktop. I set same VPN and remote desktop settings on both computers with Windows 7. I set Remote desktop to save connection credentials on both computers, but when I connect to server form one computer, it uses saved credentials and I don't get log-in prompt on server side. On the other computer with same settings using same username and password, I get log-in prompt once the remote desktop connects to server. I even tried copying the .rdp file from the computer on which everything is working fine to the other computer, but that didn't help. Note that I'm not trying to have both computers connected at the same time. What is causing this and how do I make the other computer skip windows server 2000 log-in prompt.

    Read the article

  • How do I change the default corner of the Desktop?

    - by DMA57361
    I'm using Mac OSX 10.6.8 and when I save things to the desktop, drop something there in Finder or take a screengrab with Cmd+Shift+3/4, it gets placed in the top-right corner. I tend to position my windows in a bottom-left to top-right diagonal fashion, meaning the top-right corner of the desktop is always covered, as such I'd rather that new things placed on the desktop appeared in the top-left, where there's a gap and I can get to them easily. Is it possible to adjust the "default" corner items are placed in on the Desktop? And if so, how would I go about doing so?

    Read the article

  • How to shrink the Remote Desktop Connection Bar in Windows?

    - by Some Noob Student
    Ever since FireFox 4, page tabs have been moved to the top of the window. This gets pretty irritating when surfing the net through a remote desktop session. Often when I want to switch tabs, if I accidentally move my mouse a little to much to the top, I get the remote desktop connection bar blocking the tabs instead, then I'd have to wait a few frustrating seconds before it disappears again. So, are there any methods to shrink or shorten the remote desktop connection bar or delay the appearance of it?

    Read the article

  • How to install a desktop environment onto Ubuntu Server -- but without internet access or a CDROM?

    - by James
    I am playing around with a computer which has no CDROM drive or internet access and I have installed Ubuntu Server onto it. I have that all up and running nicely but now I'd like to install Xfce, GNOME or something similar so I can load up a desktop environment from the command line if I wish. Obviously with internet access or a CDROM, this would be a simple task of using apt-get and it finding & retrieving the packages for me, I assume, but I do not have either. I do however have a USB drive and I have used Unetbootin to make it into a bootable drive with the Ubuntu Server disk image files on there. I have mounted the USB drive to /media/usb0 and tried the command "sudo apt-cdrom add -d /media/usb0" to get apt to recognise the USb drive as an "Ubuntu CD" -- a source of package files but apt-get doesn't seem to be finding Xfce.. I try "sudo apt-get install xfce" and "sudo apt-get install xfce4" but neither find the package.. I would prefer to have Xfce but GNOME would be OK too.. My question is, am I doing something wrong? I figured that the Ubuntu Server disk (or rather, my Ubuntu Server USB drive) might not have any desktop environment packages on there so I tried the Xubuntu Desktop disk too (again, from my USB drive). I tried "sudo apt-get install xubuntu-desktop" but it couldn't find the package - even though it is listed under the /casper/ directory in some MANIFEST file. Anyone see where I'm going wrong? Maybe apt-get install is looking somewhere other than my USB drive? Maybe my commands are wrong? Maybe the disks don't even have the desktop environments on!? Thanks in advance guys, any input would be much appreciated. Cheers - James

    Read the article

  • Is there a screen sharing/remote desktop app for mac that lets you use a different host screen resolution?

    - by MarqueIV
    Ok, there are tons and tons of questions about remote desktop for mac and they're all being closed as duplicates. I however am specifically looking for one that will let me use a different resolution than the host, the way you can with Remote Desktop for Windows. For instance, when I connect to my 11" Macbook Air booted into Windows7 from my quad-screen desktop, also booted into Win7 using Microsoft's Remote Desktop Client, it blanks out the screen on the notebook, then virtualizes the video across all four of my desktop's monitors at their native resolutions (2560x1600, 2 x 1920x1200 and 1600x1200) and the notebook now acts as if it has four physical monitors connected to it. All of this from a notebook that only has a 1366 x 768 native resolution. Even when running OS X on the client running RDC, while it doesn't support multi-monitors like its Win counterpart, it still lets me run at the native resolution of the client screen of 2560x1600. Again, it just blanks out the host screen while doing so. However when using Mac's screen sharing, since that is just glorified VNC, it just mirrors what's already on the host's screen, meaning it will always be a single screen with the resolution of 1366x768. This of course makes sense since VNC is a mirroring solution, not a video-virtualizing one like RDC, but it means that on my quad-monitor setup, the remote window isn't even large enough to fill up a single monitor, let alone four (unless you have a client that can scale it up, but that's video scaling. It's still only 1366x768.) So what I'm looking for is if there is a solution on the Mac that lets me do the same thing as RDC in a Win environment. Don't care if I have to pay. I'd gladly pay several hundred dollars for this. I just need that specific feature. Note: People have suggested various VNC clients, but the VNC host still runs at 1366x768 so that will not work here. Ever. Also, people have suggested Synergy/Synergy+/Teleport and such which share the keyboard and mouse, not video. Completely different animal unrelated to what I'm looking for.

    Read the article

< Previous Page | 33 34 35 36 37 38 39 40 41 42 43 44  | Next Page >