Search Results

Search found 17287 results on 692 pages for 'card game'.

Page 224/692 | < Previous Page | 220 221 222 223 224 225 226 227 228 229 230 231  | Next Page >

  • Will the netinstall (minimal) work with a USB cellular modem?

    - by Z9iT
    Installing Ubuntu minimal is extremely easy when the system has a LAN card and connected to internet via Cable & Router... I Recently came across a situation where I needed to install Ubuntu Minimal on a system which wash not having a LAN card and cable internet anywhere around... The only access to the internet was through a USB Plug n Play Data-Card Modem. May anyone suggest something so that I may install Ubuntu Minimal using USB Modem as the only source of Internet, if it's possible at all?? (This relates to Ubuntu 12.04 in specific)

    Read the article

  • How to print an Objectified Element?

    - by BeeBand
    I have xml of the format: <channel> <games> <game slot='1'> <id>Bric A Bloc</id> <title-text>BricABloc Hoorah</title-text> <link>Fruit Splat</link> </game> </games> </channel> I've parsed this xml using lxml.objectify, via: tree = objectify.parse(file) There will potentially be a number of <game>s underneath <games>. I understand that I can generate a list of <game> objects via: [ tree.games[0].game[0:4] ] My question is, what class are those objects and is there a function to print any object of whatever class these objects belong to?

    Read the article

  • How well does Intel 3000 HD work on Ubuntu?

    - by Simon
    Right now i have notebook with Nvidia 8400M GS (I know, it's not good card) and it's impossible to work normally when i'll plugin external monitor (1920x1080). Windows 7 can deal with it without problems (1440x900 on notebook + 1920x1080 external). On Ubuntu i have to choose one screen and turn off the second one. Even with only one screen Ubuntu (Unity or even Gnome3) sometimes hangs for a while, I've not found solution for this yet, but nevermind, it's probably because of my card or/and nvidia's drivers. I'm going to buy new PC, but for now only with integrated Intel 3000HD, and my question is: Should i expect similar problems with this card? Here i've found link to Intel's webpage about drivers - "only community develop them", and i'm a bit concerned. I'll use then only one monitor (the bigger one), but how well does those driver work? Are there any performance tests?

    Read the article

  • Bad texture on model with different GPU

    - by Pacha
    I have some kind of distortion on the texture of my 3D model. It works perfectly well on an AMD GPU, but when testing on a integrated Intel HD graphics card it has a weird issue. I don't have a problem with the rest of my entities as they are not scaled. The models with the problems are scaled, as my engine supports different sizes for the platforms. I am using Ogre3D as rendering engine, and GLSL as shader language. Vertex shader: #version 120 varying vec2 UV; void main() { UV = gl_MultiTexCoord0; gl_Position = gl_ModelViewProjectionMatrix * gl_Vertex; } Fragment shader: #version 120 varying vec2 UV; uniform sampler2D diffuseMap; void main(void) { gl_FragColor = texture(diffuseMap, UV); } Screenshot (the error is on the right and left side, the top and bottom part are rendered perfectly well):

    Read the article

  • Broadcom 4313 Signal strength very low on an HP Pavilion dv4

    - by thegreyspot
    I have a HP pavilion dv4 4141US laptop running fresh ubuntu 11.10. This laptop has a broadcom 4313gn chipset for the network card. The network card is using the brcmsmac driver. Which, I heard else where on askUbuntu that its the best driver to be using for this card. I can connect to my network, however I have extremely low network signal. Right now I am right in front of the router, and I only get 72%. If i go to the room directly behind the router's wall, i get no signal, where as with windows 7 I get a nice 70% percent. UPDATE: I use WEP at home and an open wireless connection at work. Both have same result. Both are completely different routers. Also you G What can i do?

    Read the article

  • Graphics problem after updating to Kernel 3.11 with Ubuntu 13.04 64bit

    - by Gaurav Sharma
    I am new to ubuntu.I have a 64 bit intel processor with ATI 6570 graphics card. Now Ubuntu 13.04 is working fine with stock kernel(3.8). But as I read somewhere that kernel 3.11 do support ATI graphics card better, I tired updating them with no success. I tried 3.11.6 3.11.4 3.11.0 but with all of them I am facing the same problem ... after installing them and restart, the screen resolution get distorted and the unity becomes too slow transparency in dash and the launcher is also lost. Now whatever little I know , this may be related with graphics diver either they are not present or the graphic card is not turned on. Can some one help with this. And yeah i tried ubuntu 13.10 that worked fine but it has some bugs. and yeah please pardon me for my bad English.

    Read the article

  • 3.5.0-18 kernel update broke BCM43228

    - by ignition
    After updating the kernel to 3.5.0-18 the Broadcom BCM43228 wireless card of my Acer Aspire One 756 has stopped working properly. The card often refuses to connect to my AP, and if it connects it behaves very unstable. I have already reinstalled the drivers, as recommended by googling around. As far as I can tell the only driver that supports my card is the bcmwl package. I found one solution saying I should install broadcom-sta using module assistant, however I could not get this to install and apparently broadcom-sta and bcmwl are effectively the same driver? Is anyone experiencing similar issues? Is it possible to just downgrade to the previous kernel version?

    Read the article

  • Disabled WiFi but it keeps re-enabling itself; want it to stay disabled

    - by Mike Uresti
    I have an Alienware X51 desktop with a wireless card, I have it connected via ethernet and want to permanently disable WiFi from ever using itself. When running Windows, I go to device manager and rightclick-disable the Wireless card and that takes care of my problem. Is there a similar method using Ubuntu 14.04? For now, I click on the network icon in the system tray and uncheck 'Enable wifi' but it seems to turn itself back on as I use my computer, I think I have the incident isolated to something with Steam. It seems like after I start the Steam client it always is re-enabled. Hopeful answer: Just disable/remove/block wireless card from ever working.

    Read the article

  • LWJGL GL_QUADS texture artifact

    - by Dajgoro Labinac
    I managed to get working lwjgl in Java, and i loaded a test image(tv test card), but i keep getting weird artifacts outside the image. Image link: http://tinypic.com/r/vhv9g/6 Code: glBegin(GL_QUADS); glTexCoord2f(0, 0); glVertex2i(10, 10); glTexCoord2f(1, 0); glVertex2i(500, 10); glTexCoord2f(1, 1); glVertex2i(500, 500); glTexCoord2f(0, 1); glVertex2i(10, 500); glEnd(); What could be the cause?

    Read the article

  • Broadcom bcm4311 rev2 Full disabled on battery

    - by Antonio BG
    Ok this is my problem... My problem begins while ago when I installed ubuntu 10.10. (is the same with 11.04,11.10...) My wireless card is disabled if the power cord is unplugged, and the only way to fix this is rebooting my pc WITH the power cord plugged and if I use lspci command, it doesn't show me the wireless card. is like if I've removed the wireless board from my laptop port. I've searched a lot of solutions for bcmw4311 issues, but none of them fix my problem because when apply one of the fixes, if I unplug the cord, my wireless card it just gone away... I have dual boot on my pc(w7,ubuntu), I've tried with full installation just to see if this is the problem, but not, is the same. On windows all work correctly with and without power cord.. so this is not a hardware problem, is more like a kernel problem. anyway I can use wireless with power cord plugged, but c'mon this is a laptop not a desktop pc XD haha any help is welcome

    Read the article

  • GameKit server disconnect makes session invalid

    - by Markus
    Hi everyone, I'm trying to build a 4 player multiplayer game using GameKit. I tried many things and I'm currently using client/server mode, meaning that I have 1 session in GKSessionModeServer and 3 other sessions connecting to this one in GKSessionModeClient. My goal is, that when a user leaves the game or gets disconnected for whatever reason, that the others can continue their game. So far this works if any of the Clients leave the game, but if the Sevrer leaves the game, send and receive calls are not reaching the other peers anymore. I tried the same with 4 peers in GKSessionModePeer with the same result (where the player who accepts connections cannot quit). Does anyone had better success in making this work? Any help is highly appreciated. Thanks!

    Read the article

  • On-board HDMI Audio

    - by TrackStar
    I have tried searching for answers to this issue but have had no luck. My HDMI display is working but I have no audio from the HDMI .There is no option for HDMI under sound preferences hardware or output. Purging and re-installing pulse audio doesn't change anything. I am using the on-board HDMI port on my motherboard, I do not have a PCI graphics card installed. CPU: i7-3770 Mobo: Asus Sabertooth z77 Kernel: 2.6.39-3-bb03 running aplay -l returns ** List of PLAYBACK Hardware Devices ** card 0: Intel [HDA Intel], device 0: ALC892 Analog [ALC892 Analog] Subdevices: 1/1 Subdevice #0: subdevice #0 card 0: Intel [HDA Intel], device 1: ALC892 Digital [ALC892 Digital] Subdevices: 1/1 Subdevice #0: subdevice #0 Seems as if the system doesn't even recognize the possibilty of HDMI audio output. Any ideas?

    Read the article

  • Load old kernel module while keeping current

    - by legion
    So I was wondering if there is anyway that I would be able to load an old iwlwifi from a 2.6 kernel while still using the 3.2 kernel and if there was a guide to doing so? I need to be able to do so because my wireless card is.... well the new kernels have broken all support for my card, which is a Intel Ultimate-N 6300. I'm extremely upset with how badly they broke support for this card to the point where I can't connect to any router that isn't wifi N enabled and even then only a select few routers work, it amazes me as to how awful it is and whats more amazing is how well it worked in 2.6 kernels. I've tried all the work arounds that have been discussed to try and fix this, and the only thing I have found that works is running a 2.6 kernel. However I would like to be able to keep my 3.2 kernel and just load up an old version of iwlwifi that actually supports it, is there anyway to do this?

    Read the article

  • Asus A8N-SLI soundcard/headphone not recognized/working

    - by Remon
    I have an Asus A8N-SLI with built-in sound-card which in the front has headset connection. Alsamixer V1.0.25 says Card: HD-Audio Generic Chip:ATI R6xx HDMI. Only option it gives is S/PDIF nothing else. I can only select the Generic sound-card others are not available. It seems it is not being recognized by Ubuntu. My technical skills are limited and I have searched for similar problems but could not find a solution. Any help in how to proceed forward is therefore greatly appreciated. I also have a windows focus problem with Gnome3 and Unity for which I will make a new question. Between those two problems it works like a charm as always. Though I'm only going to upgrade the laptop to 12.04 if I have the desktop running smoothly.

    Read the article

  • Shut down Netbook by closing lid - ubunti 13.10

    - by The Liquidator
    My wife has an Acer Aspire One (with the SSD Card and an SD expansion card). It has never been able to hibernate or suspend, always creating errors and occasionally trashing the data on the SD card, which is the home partition so it's unfortunate. To get around the problem I have set all exit methods to produce shutdown - she tends to simply shut the lid. I'm aware the default behaviour has for some time been to suspend, but I've got round that using the gnome tweak tool. However, I've just installed 13.10 and whilst I have installed the gnome tweak tool and set it to shut down the system appears to be ignoring/bypassing the setting, electing to suspend when the lid is shut. Can anyone tell me how to fix it please? I'm quite happy to get my hands dirty with the command line.

    Read the article

  • Should I switch my graphics mode in the BIOS to avoid using Bumblebee?

    - by Fawkes5
    I have just purchased a Acer 3830TG, the timeline X series. To my surprise I found out that there is no first-party support for nvidia optimus for linux. Bumblebee works great, but the battery life from the graphics card always running is not so great. I don't use linux for games so i don't really need the graphics card on, I have Windows for that. In my bios, I have the ability to change my graphics mode from switchable to integrated. If I do this, reinstall ubuntu, what will happen? Will my nvidia card just turn off? Will everything work properly, as if i'm not running an optimus laptop? Is this recommended as opposed to dealing with bumblebee? What is the best thing I could do?

    Read the article

  • How many textures can usually I bind at once?

    - by Avi
    I'm developing a game engine, and it's only going to work on modern (Shader model 4+) hardware. I figure that, by the time I'm done with it, that won't be such an unreasonable requirement. My question is: how many textures can I bind at once on a modern graphics card? 16 would be sufficient. Can I expect most modern graphics cards to support that amount? My GTX 460 appears to support 32, but I have no idea if that's representative of most modern video cards.

    Read the article

  • Object-oriented Programming - need your help

    - by wanderameise
    hey folks, I try to realize a little game project to dive deeper into OO programming (winforms c++/cli). I already started coding but now I´d like to make a re-design. For the beginning the game should consist of four parts like game-engine, user interface, highscore and playground. Heres a little (non-UML-conform) class diagramm to visualize my purposes http://i.imgur.com/lmpwj.png Would this be the right way? In my eyes the game engine is responsible to control the game sequences (state machine?) and exchanges information betweens all other classes. I appreciate any help!

    Read the article

  • Audio output and input stopped working after the last update

    - by renatov
    I'm using Ubuntu 14.04 and everything was perfect until todays's update. Now my audio output (speakers) and input (microphone) stopped working. I guess it's a driver issue, but I need help to debug this problem and to solve it. I have a Dell Inspiron 5421 notebook with an Intel audio integrated sound card: $ lspci | grep Audio 00:1b.0 Audio device: Intel Corporation 7 Series/C210 Series Chipset Family High Definition Audio Controller (rev 04) If I go to Ubuntu Settings Sound Output, it doesn't show my Intel card there anymore: The same for the Input tab, it doesn's show my Intel card there anymore: Could you please help me?

    Read the article

  • HP dv6-2119tx cant find the built in tv tuner?? Ubuntu 12.10

    - by Mat
    Installed Ubuntu 12.10 yesterday and I've been playing around trying to get all the drivers for my dv6. I have taken care of the the graphics card and the sound card which to be honest is not the greatest drivers I have found. Specs are Intel(R) Core(TM) i7 CPU Q 720 @ 1.60GHz with nvidia gt230 My problem atm is the tv tuner card! Ive searched everywhere but i cant even find something to start on. I guess I am still clueless but I really dont know what to do??? I really dont want windows anymore so if anyone could guide me Id really appreciate it. Thanks...

    Read the article

  • Radon HD 6870 compatibility with Ubuntu 11.10

    - by mhost
    Does anyone have experience with the XFX Radeon HD 6870 on Ubuntu 11.10? I've googled and the best I could find was someone getting the card to 'work' on linux. Nothing specific about ubuntu and this card. Specifically, I want to be able to use the full 1900x1200 resolution and watch 1080p movies. If yes, was it very complicated to get working? Note: I do not have the card yet, I am debating buying it, but would like to know beforehand because there is a restocking fee if I have to return it. Thanks.

    Read the article

  • Pygame's Message-multiple lines?

    - by Jam
    I am using pygame and livewires (though I don't think that part is relevant here) to create a game. I've got the game working, but I'm trying to make something akin to a title screen before the game starts. However, it doesn't recognize when I try to make a new line appear. Here is what I have: begin_message=games.Message(value=""" Destroy the Bricks!\n In this game, you control a paddle,\n controlled by your mouse,\n and attempt to destroy all the rows of bricks.\n Careful though, you only have 1 life.\n Don't mess up! The game will start in\n 5 seconds.""", size=30, x=games.screen.width/2, y=games.screen.height/2, lifetime=500, color=color.white, is_collideable=False) games.screen.add(begin_message) The message appears on the screen, but the newline doesn't happen, so I can only read the first part of the message. Is there a way to make this message actually appear, or can I not use the 'Message' for this?

    Read the article

  • How can I stop pixel seams appearing in adjacent mesh boundaries due to floating point imprecision?

    - by ufomorace
    Graphics cards are mathematically imprecise. So when some meshes are joined by their borders, the graphics card often makes mistakes and decides that some pixels at the seam represent neither object, and unwanted pixels appear. It's a natural behaviour on all graphics cards. How are such worries avoided in Pro Games? Batching? Shaders? Different tangent vectors? Merging? Overlaping seams? Dark backgrounds? Extra vertices at borders? Z precision? Camera distance tweaks? Screencap of a fix that ended up not working:

    Read the article

  • What techniques can I use to render very large numbers of objects more efficiently in OpenGL?

    - by Luke
    You can think of my application as drawing a very large ball-and-stick diagram (or graph). At times, this graph can get very large, where the number of elements even outnumbers the pixels on the screen. Currently I am simply passing all of my textures (as GL_POINTS) and lines to the graphics card using VBO's. When the number of elements outnumbers the number of pixels, is this the most efficient way to do this? Or should I do some calculations on the CPU side before handing everything over to the GPU? If it matters, I do use GL_DEPTH_TEST and GL_ALPHA_TEST. I do some alpha blending, but probably not enough to make a huge performance difference. My scene can be static at times, but the user has control over a typical arc-ball camera and can pan, rotate, or zoom. It is during these operations that performance degradation is noticeable.

    Read the article

  • How do professional games avoid showing pixel seams in adjacent mesh boundaries due to decimal imprecision?

    - by ufomorace
    Graphics cards are mathematically imprecise. So when some meshes are joined by their borders, the graphics card often makes mistakes and decides that some pixels at the seam represent neither object, and unwanted pixels appear. It's a natural behaviour on all graphics cards. How are such worries avoided in Pro Games? Batching? Shaders? Different tangent vectors? Merging? Overlaping seams? Dark backgrounds? Extra vertices at borders? Z precision? Camera distance tweaks? Screencap of a fix that ended up not working:

    Read the article

< Previous Page | 220 221 222 223 224 225 226 227 228 229 230 231  | Next Page >