Search Results

Search found 6318 results on 253 pages for 'hybrid graphics'.

Page 155/253 | < Previous Page | 151 152 153 154 155 156 157 158 159 160 161 162  | Next Page >

  • Multiplayer game communication framework for mac/ios

    - by ishaq
    (Cross post from stackoverflow) I am creating a multiplayer 2D game for Mac and iOS devices. I'll be using cocso2d for graphics/game engine, however I am largely blank on what to use for multiplayer communication. Please note that I cannot use central severs e.g. SmartFox, RedDwarf, etc since I want the players to "host" games for others and be able to play it on their LAN, VPN or my own servers. Any pointers? I checked lidgren but it's for .NET only and hence not an option for me. EDIT: just in case it wasn't clear, the messaging has to be real time hence it's probably going to be over UDP

    Read the article

  • Set nomodeset in usb installation (efi loader with iso)

    - by Garrett
    I'm working with the studyblast solution (http://studyblast.wordpress.com/2011/08/14/guide-mac-os-x-lion-how-to-boot-a-linux-live-system-from-a-usb-drive-how-to-update-any-ocz-ssds-firmware/) to creating a live usb for osx. I managed to get Ubuntu booted but the screen looks like a corrupt/glitch image. I can just about make out the clock on the bottom right and a few icons. I've been told that this is an easy fix and that nomodeset needs to be set for the graphics card but all the solutions i find are for a hard drive installed Ubuntu where you can cd in and edit the grub file - how do you do this is it's in an iso? Thanks Garrett

    Read the article

  • Licensing Theme Music from other games

    - by HS01
    As part of my game, I thought it would be fun to make a hidden level that pays tribute to Mario Bros (one of the earliest games I ever played). It would be themed in that way with 8-bit graphics and question mark blocks and completing the level would say "Thank you but the princess is in another castle" or such. For the sound track, I'm thinking of just overlaying the standard mario theme music by playing it on a virtual keyboard using a different instrument/timing or something. My question is, am I legally safe? I'm not using anyone else's actual music, I'm just playing the same tune in a different way myself. Do I have to get licensing for this?

    Read the article

  • Agile Approach for WCM

    - by cameron.f.logan
    Can anyone provide me with advice, opinions, or experience with using an agile methodology to delivery an enterprise-scale Web Content Management system (e.g., Interwoven TeamSite, Tridion)? My current opinion is that to implement a CM system there is a certain--relatively high--amount of upfront work that needs to happen to make sure the system is going to be scalable and efficient for future projects for the multi-year lifespan an WCM is expected to have. This suggests a hybrid approach at best, if not a more waterfall-like approach. I'm really interested to learn what approaches others have taken. Thanks.

    Read the article

  • Proprietary Driver missing after Kernel upgrade

    - by Chris
    I recently upgraded my kernel to the Linux 3.4 kernel. Upon doing so, I found that it zapped my proprietary driver for my graphics card. As an alternative, I had saved Catalyst 12.8, but now it will not open through Software Center. I tried several other terminal methods, but it seems that many of these postings are bunk. They only work until you're almost done and then you get errors. I just need a little help from somebody who has been there, as I am new to Ubuntu. I'm kind of afraid to reboot because this is my 3rd full load of Ubuntu. Any ideas?

    Read the article

  • HTG Projects: How to Create Your Own Custom Papercraft Toy

    - by Eric Z Goodnight
    Graphics programs aren’t simply for just editing your photos—they can have whatever fun application you can think of. For a fun, geeky project, here’s a simple papercraft toy you can make with a printer and simple household tools Latest Features How-To Geek ETC How To Boot 10 Different Live CDs From 1 USB Flash Drive The 20 Best How-To Geek Linux Articles of 2010 The 50 Best How-To Geek Windows Articles of 2010 The 20 Best How-To Geek Explainer Topics for 2010 How to Disable Caps Lock Key in Windows 7 or Vista How to Use the Avira Rescue CD to Clean Your Infected PC The Deep – Awesome Use of Metal Objects as Deep Sea Creatures [Video] Convert or View Documents Online Easily with Zoho, No Account Required Build a Floor Scrubbing Robot out of Computer Fans and a Frisbee Serene Blue Windows Wallpaper for Your Desktop 2011 International Space Station Calendar Available for Download (Free) Ultimate Elimination – Lego Black Ops [Video]

    Read the article

  • Flash player (in Firefox) is unstable in 14.04

    - by henry
    For 13.10, I used this guide to get Intel iGPU-supported playback of flash videos in Firefox: http://www.webupd8.org/2013/09/adobe-flash-player-hardware.html All in all, Intel graphics acceleration for flash content was working just fine in 13.10, at least on my machine (with Xubuntu). I basically followed the guide to do the same for 14.04, but about 2-3 weeks ago flash videos started acting up. Whenever I have one instance of flash paused or stopped in one tab and go to the next tab to continue some song on Soundcloud or something, it's game over. I'm clueless as to what I can do here. Did anything change in the meantime?

    Read the article

  • How can I get nvidia-settings to work?

    - by Palace Chan
    I have an ASUS N56VM-TB71 with a GeForce 630M graphics card. I have ubuntu 12.04 and when I vga plugged in a dell monitor, i am able to use it. When I then use my laptops hdmi port to plug in my second monitor, nothing happens, so I went to nvidia-settings and I got the error: You do not appear to be using the NVIDIA X Driver, please run nvidia-xconfig as root and restart the X server. I ran this command and rebooted but I still cannot access nvidia-settings and I keep getting that error even though my /etc/X11/xconf.org is the one that got generated by running that command! What can I do to get around that and use my second monitor as well?

    Read the article

  • What is a simple way to combine two Emacs major modes, or to change an existing mode?

    - by Winston C. Yang
    In Emacs, I'm working with a file that is a hybrid of two languages. Question 1: Is there a simple way to write a major mode file that combines two major modes? Details: The language is called "brew" (not the "BREW" of "Binary Runtime Environment for Wireless"). brew is made up of the languages R and Latex, whose modes are R-mode and latex-mode. The R code appears between the tags <% and %. Everything else is Latex. How can I write a brew-mode.el file? (Or is one already available?) One idea, which I got from this posting, is to use Latex mode, and treat the code of the form <% ... % as a comment. Question 2: How do you change the .emacs file or the latex.el file to have Latex mode treat code of the form <% ... % as a comment?

    Read the article

  • How To Use DosBox for Windows 7

    While many modern games use every bit of technology available to give gamers the latest in graphics sound and advanced gameplay some just cannot duplicate the fun that old games used to offer. A lot of fun yet old DOS games will not run on modern computers or operating systems like Windows 7 keeping you from experiencing nostalgic gaming bliss. Thanks to an emulator called DOSBox however you can enjoy many old DOS games on your Windows 7 machines and it does not matter if you are running the 32-bit or 64-bit version of the operating system.... Comcast? Business Class - Official Site Learn About Comcast Small Business Services. Best in Phone, TV & Internet.

    Read the article

  • PlayOnLinux Unable to find 32bit opengl libraries Dual ATI Videocards

    - by Rodolfo Pires
    Im curently running Ubuntu 12.04 LTS 64Bit So, i installed league of legends fine the first time with the opensource ATI Drivers provided by ubuntu itself with no issues at all, but it runs so slow ... max 20fps because those drivers dont fully support my Dual Graphic cards Than i restored system and i installed the Linux Version of the Proper ATI Drivers from the AMD Website wich supports my APU AMD-A8-4500M with the AMD Radeon 7640G + 7670M Graphics Cards enabling me full performance from my system .. Problem is, to run League of Legends i need a 32bit opengl library, and the driver, automaticly detects a 64bits Linux install and loads the 64bit libraries but not the 32 ones . i need some kind of command, to force the 32bit libraries to load, or to make League of Legends run on the 64 ones .. Im kinda noob to ubuntu .. i installed the 32 bits ones trough terminal and still doesnt work idk why, maybe the driver doesnt want to load them .. plzz help me on this, i dont want to go back to windows just to play league since im noob idk what more details to post here so plz tell me what do you need

    Read the article

  • Very High CPU usage (100%) from just browsing the Web

    - by cole
    I tested on Firefox and Chromioum. Im at 100% while loading pages which causes them to load slow and when I dont have a application running Im at 40% CPU (At least) Everything is slow basically. Im also already on Ubuntu Classic so im not using Unity. Should I go to 10.04? is that more stable? On windows this wasnt an issue. I have a Dual Boot with XP and a 2.4Ghz Intel Celeron with 768MB RAM and an Nvidia 6200 Graphics card. I heard 10.04 was the most stable. any suggestions?

    Read the article

  • KVM switching from lower resolution system resets Ubuntu high resolution

    - by Ed Manet
    I'm running 12.04 desktop on my main desktop and it's hooked to a KVM (IOGear miniview) that shares the peripherals with a SLES 11 machine. The SLES 11 machine can't get the same resolution as the Ubuntu machine because of different graphics hardware. If I switch from Ubuntu to SLES and stay there too long, when I switch back to Ubuntu the screen resolution on Ubuntu is reset to the same as SLES. I can get it back easily just opening the Displays configuration; it immediately resets to the high resolution as soon as the Displays window opens. But all my open windows have been maximized and it's a P.I.T.A. having to resize them all again. How do I get it to just stay at the high resolution between switching between systems? Is there a setting in the Xorg conf file I need to set?

    Read the article

  • Unity poor performance after upgrading to 13.10

    - by Juzer Ali
    I upgraded recently from 13.04 to 13.10. I have Dell Inspiron 4GB RAM, Intel Core i3. Login takes a lot of time and all the graphics are very slow. Alt + Tab takes a lot of time (seconds). After reading some threads, I also removed Cinnamon desktop, but it didn't change anything. I had to ultimately fall back to Gnome Flashback (no effects). Gnome flashback works fine without any issues. Can anyone please help me get back to Unity, else I will have a very hard time running using Gnome.

    Read the article

  • Right tool for the job(html5 vs silverlight)

    - by Jargo
    I have been tasked to develop a web based offer request application for bathrooms and I am asking opinions on which tool should I choose. I have narrowed options down to two (html5 and silverlight). Most of the application consists of basic UI elements that could be achieved with html4. However most challenging part is the design view for the bathroom. User first defines size of the bathroom by giving dimensions. When the dimensions are defined the application renders an empty bathroom floorplan and user can start dragging&dropping furniture on to it. Some furniture have rules where they can be dragged. For example sinks can only be dragged next to a wall. I know that this functionality can be achieved with silverlight but what about html5 ? I know that Canvas element can be used to draw graphics, but can it render the bathroom smoothly as user drags a piece furniture on it?

    Read the article

  • Switching between levels, re-initialize existing structure or create new one?

    - by Martino Wullems
    This is something I've been wondering for quite a while. When building games that exist out of multiple levels (platformers, shmups etc) what is the prefered method to switch between the levels? Let's say we have a level class that does the following: Load data for the level design (tiles), enemies, graphics etc. Setup all these elements in their appriopate locations and display them Start physics and game logic I'm stuck between the following 2 methods: 1: Throw away everything in the level class and make a new one, we have to load an entirely new level anyway! 2: pause the game logic and physics, unload all currents assets, then re-initialize those components with the level data for the new level. They both have their pros and cons. Method 1 is alot easier and seems to make sense since we have to redo everything anyway. But method 2 allows you to re-use exisiting elements which might save resources and allows for a smoother transfer to the new level.

    Read the article

  • Install AMDCONFIG on Ubuntu driver

    - by Nick Bailuc
    In 12.04 I used the official driver downloaded from amd.com which came with amdconfig but now in 14.04 the official driver is buggy so I just use the Ubuntu Official Drivers which works even better because they beefed up the original driver. The Ubuntu driver doesnt come with the terminal command amdconfig which allowed me to tweak/overclock my graphics card. How can I install it without having to install the original AMD driver? Additional Information: -I only use x.org drivers becuase it's opensource therefore more stable rather than the proprietary fglrx driver -I do not use procrams like amdoverdrivectrl or atioverclock because they are not as stable and advanced as the terminal command

    Read the article

  • After upgrading to 13.04 Unity interface is not showing

    - by ?????? ?????
    I've upgraded to Raring last night. The upgrade itself went okay, no errors. But when I rebooted the computer afterwards and logged in to my Unity session, all I could see was the Desktop background (together with Desktop icons), and no Unity interface. The Super button shortcut wasn't showing the Dash, there was no top panel etc. Please see the screenshot. As a hint, I'm suspecting it's got something to do with my switchable graphics. I'm running Ubuntu on Acer Aspire AS5830TG with nVidia GT540M and an Intel integrated card. In 12.10 I was using Bumblebee to manage the graphic card switching. During the upgrade I saw something related to nvidia had to be uninstalled, but didn't pay much attention to it. I can't be sure if it has anything to do with my problem though. What could possibly go wrong?

    Read the article

  • Multiple Monitors using nvidia-prime or bumblebee on Ubuntu 13.10

    - by user205626
    I've been unable to get multiple monitors to work with Ubuntu 13.10 using nvidia-prime or bumblebee. Could someone point me in the right direction? With nvidia-prime, I've tried the xorg.conf here http://us.download.nvidia.com/XFree86/Linux-x86/319.12/README/randr14.html, but I boot into "low graphics" mode and have to revert to get a desktop back. Any suggestions would be appreciated. Thanks. Edit: I've given up on nvidia-prime; I missed the fact that it never turns off the discrete card... So, I'm back to trying to get VIRTUAL displays working with Bumblebee.

    Read the article

  • Purple start screen - no splash screen

    - by Peter
    After installing Ubuntu 11.10 on a new computer everything worked fine for a few weeks and now suddenly when starting up Ubuntu I get a blank purple screen with no "ubuntu" splash and the 5 dots but the system does start up. What has happened to loose the ubuntu splash screen? Also, I've checked in the grub startup and it has "quiet splash" in the linux line. I have an ATI radeon 5450 graphics card which needed nomodeset to be added to the grub line when booting for the first time off the install CD but now this line has disappeared from the grub startup. Can anyone help. Thanks

    Read the article

  • Complete Guide/Tutorials on LWJGL?

    - by user43353
    Dont get me wrong, I finished these tutorials on http://lwjgl.org/wiki/index.php?title=Main_Page. I finished The Basics section, OpenGL 3.2 and newer section, and I looked at the Example Code section. They were great tutorials, and I have looked at the external tutorials as well. I don't know where to go from here, and OpenGL is not my strong point. Some one suggested Learning Modern 3D Graphics Programming, and I didnt learn much. I looked at the port to LWJGL, but the book was on C and I couldn't really understand what the OpenGL meant. I am trying to learn 2D gaming, not 3D. Maybe later. Is there any tutorials that aren't C/C++ heavy and teach you 2D OpenGL?

    Read the article

  • Horible lag on Xubuntu media center setup

    - by William
    I am setting up a media center running Xubuntu 11.10 as a media center. It is using the XFCE desktop enviroment. I have the NVIDIA drivers installed from the Additional Drivers program, and uninstalled all but the esennitial software. The computer is a Dell Dimension 3000. You can view detailed technical specifications here. The graphics card installed is a Geforce 8400 GS. I have installed all updates through Update Manager. What can I do to fix the horrible lag? It makes watching TV impossible.I have uploaded two videos of the lag, here and here. The second one does not have sound because the sound stops playing on some channels that lag a lot. The TV card is a Hauppage 950q USB, and the TV recording program is Kaffiene. Please help me with this, I am trying to do this to show how great Ubuntu is to the rest of the family.

    Read the article

  • GNOME 3.4 released, with smooth & fast magnification

    - by Peter Korn
    The GNOME community released GNOME 3.4 today. This release contains several new accessibility features, along with a new set of custom high-contrast icons which improve the user experience for users needing improved contrast. This release also makes available the AEGIS-funded GNOME Shell Magnifier. This magnifier leverages the powerful graphics functionality built into all modern video cards for smooth and fast magnification in GNOME. You can watch a video of that magnifier (with the previous version of the preference dialog), which shows all of the features now available in GNOME 3.4. This includes full/partial screen magnification, a magnifier lens, full or partial mouse cross hairs with translucency, and several mouse tracking modes. Future improvements planned for GNOME 3.6 include focus & caret tracking, and a variety of color/contrast controls.

    Read the article

  • How can I reconfigure the nvidia proprietary drivers from the command line (ssh)?

    - by Mathieu Pagé
    I have a linux HTPC (running XBMC) in my living room. This morning I ssh'ed into the machine and did upgrade it to 10.10. When it finaly resarted it says something about running in low quality graphics and eventually returned to a command line login prompt. I ssh'ed in again and did a sudo reboot now. When it came back on this time the image is rapidly scrolling from the top to the bottom of the screen. I guess the installed driver doesn't quite work with the S-Video port on which the TV is connected. previously it was working right with the nvidia proprietary drivers. How can I install thoses without using the GUI tool that comes with Ubuntu?

    Read the article

  • OpenGL 3 and the Radeon HD 4850x2

    - by rotard
    A while ago, I picked up a copy of the OpenGL SuperBible fifth edition and slowly and painfully started teaching myself OpenGL the 3.3 way, after having been used to the 1.0 way from school way back when. Making things more challenging, I am primarily a .NET developer, so I was working in Mono with the OpenTK OpenGL wrapper. On my laptop, I put together a program that let the user walk around a simple landscape using a couple shaders that implemented per-vertex coloring and lighting and texture mapping. Everything was working brilliantly until I ran the same program on my desktop. Disaster! Nothing would render! I have chopped my program down to the point where the camera sits near the origin, pointing at the origin, and renders a square (technically, a triangle fan). The quad renders perfectly on my laptop, coloring, lighting, texturing and all, but the desktop renders a small distorted non-square quadrilateral that is colored incorrectly, not affected by the lights, and not textured. I suspect the graphics card is at fault, because I get the same result whether I am booted into Ubuntu 10.10 or Win XP. I did find that if I pare the vertex shader down to ONLY outputting the positional data and the fragment shader to ONLY outputting a solid color (white) the quad renders correctly. But as SOON as I start passing in color data (whether or not I use it in the fragment shader) the output from the vertex shader is distorted again. The shaders follow. I left the pre-existing code in, but commented out so you can get an idea what I was trying to do. I'm a noob at glsl so the code could probably be a lot better. My laptop is an old lenovo T61p with a Centrino (Core 2) Duo and an nVidia Quadro graphics card running Ubuntu 10.10 My desktop has an i7 with a Radeon HD 4850 x2 (single card, dual GPU) from Saphire dual booting into Ubuntu 10.10 and Windows XP. The problem occurs in both XP and Ubuntu. Can anyone see something wrong that I am missing? What is "special" about my HD 4850x2? string vertexShaderSource = @" #version 330 precision highp float; uniform mat4 projection_matrix; uniform mat4 modelview_matrix; //uniform mat4 normal_matrix; //uniform mat4 cmv_matrix; //Camera modelview. Light sources are transformed by this matrix. //uniform vec3 ambient_color; //uniform vec3 diffuse_color; //uniform vec3 diffuse_direction; in vec4 in_position; in vec4 in_color; //in vec3 in_normal; //in vec3 in_tex_coords; out vec4 varyingColor; //out vec3 varyingTexCoords; void main(void) { //Get surface normal in eye coordinates //vec4 vEyeNormal = normal_matrix * vec4(in_normal, 0); //Get vertex position in eye coordinates //vec4 vPosition4 = modelview_matrix * vec4(in_position, 0); //vec3 vPosition3 = vPosition4.xyz / vPosition4.w; //Get vector to light source in eye coordinates //vec3 lightVecNormalized = normalize(diffuse_direction); //vec3 vLightDir = normalize((cmv_matrix * vec4(lightVecNormalized, 0)).xyz); //Dot product gives us diffuse intensity //float diff = max(0.0, dot(vEyeNormal.xyz, vLightDir.xyz)); //Multiply intensity by diffuse color, force alpha to 1.0 //varyingColor.xyz = in_color * diff * diffuse_color.xyz; varyingColor = in_color; //varyingTexCoords = in_tex_coords; gl_Position = projection_matrix * modelview_matrix * in_position; }"; string fragmentShaderSource = @" #version 330 //#extension GL_EXT_gpu_shader4 : enable precision highp float; //uniform sampler2DArray colorMap; //in vec4 varyingColor; //in vec3 varyingTexCoords; out vec4 out_frag_color; void main(void) { out_frag_color = vec4(1,1,1,1); //out_frag_color = varyingColor; //out_frag_color = vec4(varyingColor, 1) * texture(colorMap, varyingTexCoords.st); //out_frag_color = vec4(varyingColor, 1) * texture(colorMap, vec3(varyingTexCoords.st, 0)); //out_frag_color = vec4(varyingColor, 1) * texture2DArray(colorMap, varyingTexCoords); }"; Note that in this code the color data is accepted but not actually used. The geometry is outputted the same (wrong) whether the fragment shader uses varyingColor or not. Only if I comment out the line varyingColor = in_color; does the geometry output correctly. Originally the shaders took in vec3 inputs, I only modified them to take vec4s while troubleshooting.

    Read the article

< Previous Page | 151 152 153 154 155 156 157 158 159 160 161 162  | Next Page >