Search Results

Search found 8831 results on 354 pages for 'intel graphics'.

Page 211/354 | < Previous Page | 207 208 209 210 211 212 213 214 215 216 217 218  | Next Page >

  • Black screen after upgrading from 13.04 to 13.10

    - by Harri
    Just upgraded from 13.04 to 13.10 and all I got was a black screen. The hardware I'm running is Asus Zenbook UX31A (Intel GPU). I do hear that the login screen drums do play, so the system does boot to login screen. When I try to boot using kernel 3.11.0-12 recovery mode, it tells me "initctl: event failed". Then if I go on an press ctrl+alt+f2, log in and command startx, it dies because "Fatal server error: no screens found". Here are some logs from /var/log/Xorg.0.log http://pastebin.com/ZQasUKJx Kernel 3.8.0-31 work ok, as did things before the upgrade.

    Read the article

  • Analog and digital audio output at the same time

    - by wim
    My speakers use a digital input, but my headphones use an analog input. I have them both plugged in, and when I want to use the headphones I just turn off the speakers and switch on the headphones. I know that simultaneous output on digital and analog is supported by the hardware, because it worked fine in Windows XP. But on Ubuntu, I seem to only get one at a time, depending on which setting is selected in the combo box located at System -> Preferences -> Sound -> Hardware. How can I get simultaneous analog and digital output without having to switch the profile every time? I'm on Ubuntu 11.04 and it's an HDA Intel chip.

    Read the article

  • Picking up radio station when connecting monitor on latop

    - by Goddard
    I have a laptop Asus G73JW ROG-A3B7M Intel Core i7 || 16GB RAM || 802.11n+BT || 2 x 500GB Hybrid Drives || Nvidia GTX 460M and I am picking up some staticy radio station when connecting my monitor over RGB connector. My radio system is connected via my laptops headphone jack. If I disconnect my extra monitor the radio station static disappears. I am not sure what this is related to or even how to debug such an issue. Any help is appreciated. The monitor is a Acer P216HL and the stero is a TeAC nxt

    Read the article

  • After upgrading to 13.04 Unity interface is not showing

    - by ?????? ?????
    I've upgraded to Raring last night. The upgrade itself went okay, no errors. But when I rebooted the computer afterwards and logged in to my Unity session, all I could see was the Desktop background (together with Desktop icons), and no Unity interface. The Super button shortcut wasn't showing the Dash, there was no top panel etc. Please see the screenshot. As a hint, I'm suspecting it's got something to do with my switchable graphics. I'm running Ubuntu on Acer Aspire AS5830TG with nVidia GT540M and an Intel integrated card. In 12.10 I was using Bumblebee to manage the graphic card switching. During the upgrade I saw something related to nvidia had to be uninstalled, but didn't pay much attention to it. I can't be sure if it has anything to do with my problem though. What could possibly go wrong?

    Read the article

  • HP ENVY 4-Sleekbook or Samsung Series 5 NP530U3B Ultrabook?

    - by Sam
    I am a high school student and I need a laptop within the budget of 650$. I usually have a browser, microsoft office, music, and possibly a movie or something open at once. Will the HP ENVY 4's Intel Core i3 processor be enough to handle this or would I have to get the Samsung series 5 13 inch ultrabook to get this job done? I really like the look of the HP ENVY 4, but I also want a laptop that will be quick enough to handle my needs. PLEASE HELP!

    Read the article

  • Turning off the backlight on the console

    - by Albertas Agejevas
    I'm using an IBM Thinkpad X21 as a 3G router. It's running a server install of lucid. The laptop has the lid up most of the time. There is no X, just a plain console. The colsole blanks itself, but the LCD backlight is always on. The graphics chip is ATI Technologies Inc Rage Mobility P/M AGP 2x (rev 64). The following modules related to the framebuffer are now loaded: $ lsmod | grep fb fbcon 35102 71 tileblit 2031 1 fbcon font 7557 1 fbcon bitblit 4707 1 fbcon vga16fb 11385 1 vgastate 8961 1 vga16fb Can I make the LCD backlight turn off when the colsole is unused and have it turn on automatically when needed?

    Read the article

  • Multiple Monitors using nvidia-prime or bumblebee on Ubuntu 13.10

    - by user205626
    I've been unable to get multiple monitors to work with Ubuntu 13.10 using nvidia-prime or bumblebee. Could someone point me in the right direction? With nvidia-prime, I've tried the xorg.conf here http://us.download.nvidia.com/XFree86/Linux-x86/319.12/README/randr14.html, but I boot into "low graphics" mode and have to revert to get a desktop back. Any suggestions would be appreciated. Thanks. Edit: I've given up on nvidia-prime; I missed the fact that it never turns off the discrete card... So, I'm back to trying to get VIRTUAL displays working with Bumblebee.

    Read the article

  • OpenGL 3 and the Radeon HD 4850x2

    - by rotard
    A while ago, I picked up a copy of the OpenGL SuperBible fifth edition and slowly and painfully started teaching myself OpenGL the 3.3 way, after having been used to the 1.0 way from school way back when. Making things more challenging, I am primarily a .NET developer, so I was working in Mono with the OpenTK OpenGL wrapper. On my laptop, I put together a program that let the user walk around a simple landscape using a couple shaders that implemented per-vertex coloring and lighting and texture mapping. Everything was working brilliantly until I ran the same program on my desktop. Disaster! Nothing would render! I have chopped my program down to the point where the camera sits near the origin, pointing at the origin, and renders a square (technically, a triangle fan). The quad renders perfectly on my laptop, coloring, lighting, texturing and all, but the desktop renders a small distorted non-square quadrilateral that is colored incorrectly, not affected by the lights, and not textured. I suspect the graphics card is at fault, because I get the same result whether I am booted into Ubuntu 10.10 or Win XP. I did find that if I pare the vertex shader down to ONLY outputting the positional data and the fragment shader to ONLY outputting a solid color (white) the quad renders correctly. But as SOON as I start passing in color data (whether or not I use it in the fragment shader) the output from the vertex shader is distorted again. The shaders follow. I left the pre-existing code in, but commented out so you can get an idea what I was trying to do. I'm a noob at glsl so the code could probably be a lot better. My laptop is an old lenovo T61p with a Centrino (Core 2) Duo and an nVidia Quadro graphics card running Ubuntu 10.10 My desktop has an i7 with a Radeon HD 4850 x2 (single card, dual GPU) from Saphire dual booting into Ubuntu 10.10 and Windows XP. The problem occurs in both XP and Ubuntu. Can anyone see something wrong that I am missing? What is "special" about my HD 4850x2? string vertexShaderSource = @" #version 330 precision highp float; uniform mat4 projection_matrix; uniform mat4 modelview_matrix; //uniform mat4 normal_matrix; //uniform mat4 cmv_matrix; //Camera modelview. Light sources are transformed by this matrix. //uniform vec3 ambient_color; //uniform vec3 diffuse_color; //uniform vec3 diffuse_direction; in vec4 in_position; in vec4 in_color; //in vec3 in_normal; //in vec3 in_tex_coords; out vec4 varyingColor; //out vec3 varyingTexCoords; void main(void) { //Get surface normal in eye coordinates //vec4 vEyeNormal = normal_matrix * vec4(in_normal, 0); //Get vertex position in eye coordinates //vec4 vPosition4 = modelview_matrix * vec4(in_position, 0); //vec3 vPosition3 = vPosition4.xyz / vPosition4.w; //Get vector to light source in eye coordinates //vec3 lightVecNormalized = normalize(diffuse_direction); //vec3 vLightDir = normalize((cmv_matrix * vec4(lightVecNormalized, 0)).xyz); //Dot product gives us diffuse intensity //float diff = max(0.0, dot(vEyeNormal.xyz, vLightDir.xyz)); //Multiply intensity by diffuse color, force alpha to 1.0 //varyingColor.xyz = in_color * diff * diffuse_color.xyz; varyingColor = in_color; //varyingTexCoords = in_tex_coords; gl_Position = projection_matrix * modelview_matrix * in_position; }"; string fragmentShaderSource = @" #version 330 //#extension GL_EXT_gpu_shader4 : enable precision highp float; //uniform sampler2DArray colorMap; //in vec4 varyingColor; //in vec3 varyingTexCoords; out vec4 out_frag_color; void main(void) { out_frag_color = vec4(1,1,1,1); //out_frag_color = varyingColor; //out_frag_color = vec4(varyingColor, 1) * texture(colorMap, varyingTexCoords.st); //out_frag_color = vec4(varyingColor, 1) * texture(colorMap, vec3(varyingTexCoords.st, 0)); //out_frag_color = vec4(varyingColor, 1) * texture2DArray(colorMap, varyingTexCoords); }"; Note that in this code the color data is accepted but not actually used. The geometry is outputted the same (wrong) whether the fragment shader uses varyingColor or not. Only if I comment out the line varyingColor = in_color; does the geometry output correctly. Originally the shaders took in vec3 inputs, I only modified them to take vec4s while troubleshooting.

    Read the article

  • A good way to build a game loop in OpenGL

    - by Jeff
    I'm currently beginning to learn OpenGL at school, and I've started making a simple game the other day (on my own, not for school). I'm using freeglut, and am building it in C, so for my game loop I had really just been using a function I made passed to glutIdleFunc to update all the drawing and physics in one pass. This was fine for simple animations that I didn't care too much about the frame rate, but since the game is mostly physics based, I really want to (need to) tie down how fast it's updating. So my first attempt was to have my function I pass to glutIdleFunc (myIdle()) to keep track of how much time has passed since the previous call to it, and update the physics (and currently graphics) every so many milliseconds. I used timeGetTime() to do this (by using <windows.h>). And this got me to thinking, is using the idle function really a good way of going about the game loop? My question is, what is a better way to implement the game loop in OpenGL? Should I avoid using the idle function?

    Read the article

  • Horible lag on Xubuntu media center setup

    - by William
    I am setting up a media center running Xubuntu 11.10 as a media center. It is using the XFCE desktop enviroment. I have the NVIDIA drivers installed from the Additional Drivers program, and uninstalled all but the esennitial software. The computer is a Dell Dimension 3000. You can view detailed technical specifications here. The graphics card installed is a Geforce 8400 GS. I have installed all updates through Update Manager. What can I do to fix the horrible lag? It makes watching TV impossible.I have uploaded two videos of the lag, here and here. The second one does not have sound because the sound stops playing on some channels that lag a lot. The TV card is a Hauppage 950q USB, and the TV recording program is Kaffiene. Please help me with this, I am trying to do this to show how great Ubuntu is to the rest of the family.

    Read the article

  • Audio codec consuming high battery power

    - by Vamsi Emani
    My powertop reports this for the two audio codec components. 4.85 W 100.0% Device Audio codec hwC0D3: Intel 4.85 W 100.0% Device Audio codec hwC0D0: Realtek I think 10 W for audio is too high. Can somebody please suggest me a way to reduce the power consumption? It'd be nice if someone could educate me on this, I have an idea about codecs in general but I have no clue about their internals? Why is it that these two components keep running always even when I am not listening to audio?

    Read the article

  • Does the latest version of Ubuntu (12.04) support Unity 3D?

    - by Douglas Combs
    I just installed the latest version of Ubuntu (12.04) 64bit. I am using a Radeon HD 7750 vid card. I think I have the Catalyst driver installed correctly. But when I go to system and look at the details, it shows that my graphics is VESA:01. Does this mean I it, I didn't correctly install my driver? System Specs: MB: ASUS P7P55-M CPU: Intel i5 Quad Core MEM: 4GB DD3 VC: HIS Radeon HD 7750 (1GB DDR5) Thanks for help.

    Read the article

  • Ubuntu Hangs Suddenly (Dell Latitude E5530)

    - by iFadey
    I recently bought (a month ago) Dell Latitude E5530 which comes pre-installed with Ubuntu 11.10. I removed Ubuntu 11.10 and installed 12.04 LTS right after the purchase. Everything worked out of box but occasionally Ubuntu completely hangs. The screen freezes and I can't even switch to other terminals by pressing (CTRL+ALT+F*). Whenever the screen freezes, CPU fan speed also increases. This is not happening when running particular applications. I mean it can hang without giving any reason or error displayed and while running any application. In short currently I can't able to reproduce system hang myself. I also want to mention that sometimes it never hangs complete day. Here are the specs of my laptop: Processor: Core i7-3520M CPU @ 2.90GHz Memory: 8GB HDD: 500GB, 7200rpm (Model=ST9500423AS) Graphics: Intel HD 4000 Operating System: Ubuntu 12.04 LTS (64-bit) Thanks!

    Read the article

  • window decoration command change

    - by Pedro
    I'm quite new to Ubuntu, and I don't know how to fix this: I have been experiencing the problem with window decorations that go away. I have been told that I could correct that with compiz fusion icon, just restarting the graphics whenever it happened. This solution only fixed the problem for like 30 seconds. Then I found out that the command inside the window decorator had changed, so I only had to reset it to default (/usr/bin/compiz-decorator). The only problem is that it keeps changing the command by itself once or twice a day. I can live with that, but I would like to find a more permanent solution, if there is one. I am running Ubuntu 11.10

    Read the article

  • Lenovo u400, Ubuntu 12.04. Sudden reboots / power down

    - by Victor Neo
    I've recently bought a Lenovo u400 (http://www.lenovo.com/products/us/laptop/ideapad/u-series/u400/) and installed Ubuntu 12.04 on it. Everything works fine (I've installed the proprietary ATI drivers for switchable graphics) except that I get a random power down or reboot either while surfing the web or watching videos. Monitoring the temperature with sensors, the laptop never goes into any critical zone but still may power down or reboot once in a while. Is there any logs that I should check and any particular message I should take note of? EDIT: This happens to me as well. This also happens without proprietary drivers activated.

    Read the article

  • Kernel freezes with nvidia optimus and bumblebee on 12.04

    - by piovisqui
    I have a Sony VAIO with Intel i7-3520M and NVIDIA Corporation GK107M [GeForce GT 640M LE]. I installed bumblebee and nvidia drivers. They work almost fine, but the system freezes like in a kernel panic once every day. I can play Dota 2 with primusrun on it with 20-30fps 1080p. But when I do something simple as expanding the KDE calendar widget from the bottom right corner, the system hangs a lot and windows borders get another border. The system freezes in not very intensive graphic tasks, for instance, ALT+TAB with KWin effects or pressing ALT+F2 for poping up the launcher. How can I solve this? Info: I use KDE. Couldn't paste de logs here, check at pastebin: http://pastebin.com/SBet8fRQ

    Read the article

  • How to Access Boot Options 12.04 Live USB

    - by Ryan Kampmeier
    I'm attempting to install 12.04 on my computer but booting from my USB drive results in a blank screen. This has happened since 10.04 and has always been fixed by putting nomodeset in the boot command, but now I can't access the boot options with F6 for some reason. How can I access the boot options? I booted into the live image on a different computer and it worked fine. I'm sure it's because of my Nvidia graphics. When I boot into the live image it shows a small gray box in the lower left corner of the screen and that's it. I can't edit the boot options or do anything else. Thanks in advance.

    Read the article

  • System network service uncompatable after harddrive change and format

    - by NStephenH
    After changing out the hard drive in my dell optiplex 750, and reinstalling ubuntu, my computer tells me the system network services are not compatable. Because my cd drive does not work, I installed this by placing the hard drive in another computer an dusign the mini.iso to install it. I need networking capabilities becuase I want to use this computer as a web server. The on-board ethernet card is a intel 82566 something. If you need more information, and can tell me how to get it, just ask. Please help.

    Read the article

  • AHCI Hotswap in Lucid Server

    - by chong
    Does Lucid support hotswapping of AHCI devices? Is there anything more to it than simply unmounting all filesystems on the drive to be disconnected and plugging in the new disk? 00:1f.2 SATA controller: Intel Corporation 82801IR/IO/IH (ICH9R/DO/DH) 6 port SATA AHCI Controller (rev 02) is the controller I'm working with. BIOS is set to AHCI. I also tried simply unmounting and unplugging. That seemed to work fine. Plugging in a new drive locked up the system. AHCI kernel module is also loaded. I've also looked at the Raid WIKI on kernel.org which talks about hardware requirements for AHCI HotSwapping of disks. My drives and cables all appear to meet the criteria, but most of that info pertains to the electrical side (grounding, etc...).

    Read the article

  • Input Signal Out of Range 1920x1080

    - by Zach
    I've recently installed Ubuntu 12.04 on my computer. Now, every time I boot and when I shut down, my monitor goes blank and says "Input Signal Out of Range - Change Setting to 1920x1080 60Hz." Once the computer gets to the login screen, it's okay again. This problem also happens when I try to open any 3d app. My graphics card is NVIDIA GEforce 6150 SE. I tried updating the drivers, but it broke everything and I had to reinstall Ubuntu. Any help?

    Read the article

  • Dimming the backlight is irreversible on a Samsung Q210 notebook, what do I do?

    - by user27304
    I'm new to the community, although I have been using Ubuntu since 2010. I have a Samsung Q210 notebook; Specs: Intel® Core™2 Duo CPU P8400 @ 2.26GHz × 2 4 Gigs RAM Nvidia 9200m GS (although system information in Ubuntu doesn't know) 194 GB HD OS: Ubuntu 11.10 Kernel is 3.0.0-12-generic-pae Although Samsung seems to be infamous for problems with Ubuntu, after upgrading to Oneiric, finally the FN Brightness Buttons are recognized. The only problem is, after dimming the backlight for a fixed amount of steps (3 or 4, I dare not count now because that would mean rebooting because I can't see anything), the display goes completely dark and using the FN buttons to brighten the backlight does not work anymore (before reaching that threshold, going brighter after dimming works). Now what do I do? File a bug report? If not, what then? If yes, how? Not sure... guess I should ask here first.. thanks for answering in advance.

    Read the article

  • Gnome shell crashing in 11.10 with 'purple haze' effect

    - by Andy
    I've just got Gnome with 11.10 up and running with my netbook and liked it so much I thought I'd get my parents old machine sorted with Ubuntu too. Unity works fine but when I try to switch to Gnome shell, I get problems. On login, the wallpaper appears as normal but then the colour bleeds out from the centre, leaving traces of it around the edge in a purple haze (yes, I do like Jimi Hendrix but I'm not making this up). When I go to applications, typing then pressing enter in the search field dumps me out so there's nothing but File / Edit / View etc in top left; starting a program seems to work but then there's only a white screen and no program window. Gnome classic works fine, from the limited use I've given it. I'm using an Asus desktop with ABIT motherboard, 2.6ghz with 1gb RAM; I've checked drivers and it says I'm up to date, Nvidia graphics. Anyone any ideas?

    Read the article

  • How can I reconfigure the nvidia proprietary drivers from the command line (ssh)?

    - by Mathieu Pagé
    I have a linux HTPC (running XBMC) in my living room. This morning I ssh'ed into the machine and did upgrade it to 10.10. When it finaly resarted it says something about running in low quality graphics and eventually returned to a command line login prompt. I ssh'ed in again and did a sudo reboot now. When it came back on this time the image is rapidly scrolling from the top to the bottom of the screen. I guess the installed driver doesn't quite work with the S-Video port on which the TV is connected. previously it was working right with the nvidia proprietary drivers. How can I install thoses without using the GUI tool that comes with Ubuntu?

    Read the article

  • Why are 3D game scenarios so big? and.. are them shareable? [closed]

    - by Néstor Sánchez A.
    First, I apologize if this is not the forum for these type of questions. I was wondering... 1- Why 3D games (such as GTA IV) uses so much gigabytes of space? And my half-answar was... because of their huge 3D scenarios. But wait! are not these scenarios made with vector graphics and textures? then.. why are so big??? 2- If they are so big... would really these companies make a whole new scenario for every new game they develop? that would be very inefficient. Doesn't exist "shareable" scenarios that can be used in different games and expanded to no rebuild a whole one (like "New York" or "Liberty City")???. I mean.. the scenarios is almost the same in terms of vertices, just with improved textures.

    Read the article

  • What is the practical use of IBOs / degenerate vertex in OpenGL?

    - by 0xFAIL
    Vertices in 3D models CAN get cut in the process of optimizing 3D geometry, (degenerate vertices) by 3D graphics software (Blender, ...) when exporting because they aren't needed when reusing a vertex for multiple triangles. (In the current case 3D data is exported from Blender as .ply and read by a simple application that displays the 3D model) Every vertex has a few attributes like position, color, normal, tangent,... But the data for each vertex that is cut through the vertex sharing is lost and is missing in the vertex shader. Modern shader techniques like Bump or Normal mapping require normals/tangents per vertex which are also cut. To use complex shader techniques IBOs must not be used? Or is there a way to use IBOs and retain the data per vertex that was origionally lost?

    Read the article

< Previous Page | 207 208 209 210 211 212 213 214 215 216 217 218  | Next Page >