Search Results

Search found 13860 results on 555 pages for 'core graphics'.

Page 54/555 | < Previous Page | 50 51 52 53 54 55 56 57 58 59 60 61  | Next Page >

  • Is it possible to update my graphic card drivers?

    - by madmaxpt
    This is what the lspci -v command shows in the VGA compatible controller section: 00:02.0 VGA compatible controller: Intel Corporation Atom Processor D4xx/D5xx/N4xx/N5xx Integrated Graphics Controller (prog-if 00 [VGA controller]) Subsystem: Dell Device 048e Flags: bus master, fast devsel, latency 0, IRQ 44 Memory at f0200000 (32-bit, non-prefetchable) [size=512K] I/O ports at 18d0 [size=8] Memory at d0000000 (32-bit, prefetchable) [size=256M] Memory at f0000000 (32-bit, non-prefetchable) [size=1M] Expansion ROM at <unassigned> [disabled] Capabilities: <access denied> Kernel driver in use: i915 Kernel modules: i915 i915 is only compatible with OpenGL 1.4 and cannot use shaders and I'd like to use some Open GL 2.0 functionalities. Is it possible to update the drivers or is my hardware limited to OpenGL 1.4 ?

    Read the article

  • in ubuntu 12.04 how may i know if my devices have it's driver installed??

    - by Aldo
    i have a dell N4110 laptop, and i want to know if the driver is installed and working well, something like the device manager in windows , or another way to know if a device is driverless or if the device might have a better driver, like my mousepad, in windows it have multi-touch gestures , that scrolls or zoom with two fingers (like an ipod) but in ubuntu it just works the right part as a scroll bar, so maybe it is installed one driver, but i need other one that uses well my devices. and the grphics card, i have not idea if it is well installed or isn't. i have a Intel 3000hd graphics card. thank you for your time. have a nice day people! =D

    Read the article

  • How to install & upgarde my Video Drive in presice?

    - by user71382
    Greeting all out here, So one helpful folk during a troubleshoot on VirtualBox told me that my graphic driver is too old. Code: 00:00:02.714 OpenGL Info: Render SPU: GL_VENDOR: Tungsten Graphics, Inc 00:00:02.714 OpenGL Info: Render SPU: GL_RENDERER: Mesa DRI Mobile Intel® GM45 Express Chipset x86/MMX/SSE2 00:00:02.714 OpenGL Info: Render SPU: GL_VERSION: 2.1 Mesa 8.0.2 In try to shot the problem, I got at http://intellinuxgraphics.org/2012.02.html But now I don't know how to proceed & what to download & how to install the latest driver for my video card ????????????/ Any help on the matter is appreciated. Thank you,Have a nice day, Mrinal

    Read the article

  • Oneiric updates cause "Window Creation Error" in Second Life client

    - by Yuttadhammo
    With the latest Oneiric updates, I've been unable to use Second Life, a virtual reality simulator, in Ubuntu. I assume it has something to do with the NVidia driver, since the error in the console says: WARNING: createWindow: LLWindowManager::create() : Error creating window. WARNING: LLViewerWindow: Failed to create window, to be shutting Down, be sure your graphics driver is updated I guess I could file a bug on Launchpad, but I'm not sure where to file it. Does anyone have any insight into what may be causing this problem? It was working fine a week ago, and older viewers (Imprudence viewer anyway) still work fine.

    Read the article

  • How to force GNOME panels et. al. to display on a different monitor without mirroring?

    - by GrueKun
    So, I recently purchased a new 23" monitor for my PC. However, I can't use it with the PC currently as I am waiting on a replacement heatsink. In the mean time, I wanted to use it with my Dell laptop. I hooked it up to the VGA port, and it seems to be working properly. However, I wanted to know if there was a way I could move all of the main display elements over to the attached monitor? I wanted to shut the LCD panel off on the laptop and hook it up like a desktop. Relevant specs: Ubuntu 10.10 x64 Intel graphics chipset The attached monitor is currently set as the default monitor. Any suggestions are welcome. :)

    Read the article

  • Problem installing libva (VAAPI), vainfo fails

    - by satuon
    I'm following a wiki on how to make VLC use the GPU in the Core i3 integrated graphics chipset, but I'm stuck at one of the first steps, installing VAAPI. I installed libva1 and vainfo by using "sudo apt-get install libva1 vainfo", but when I run vainfo it says libva: libva version 0.32.0 libva: va_getDriverName() returns 0 libva: Trying to open /usr/lib/dri/i965_drv_video.so libva: va_openDriver() returns -1 vaInitialize failed with error code -1 (unknown libva error),exit It should say VAProfileH264High : VAEntrypointVLD VAProfileVC1Advanced : VAEntrypointVLD according to the wiki. /usr/lib/dri/i965_drv_video.so doesn't exist in my system, but I installed libva1 with apt-get.

    Read the article

  • Which is the best low/mid level GPU in terms of compatibility for 12.04 and 12.10?

    - by Ghost
    I'm tired of dealing with the disaster that is catalyst on a 4xxx IGP and I want to buy some discrete graphics that support both Unity and alternative environments like GNOME3 and Mate-Cinnamon without getting sluggish. Besides that, 1080p video and some light 3D gaming. Again nothing crazy, I'm not a designer and I wont be using stuff like Blender. One thing I want to make clear is that I want a GPU that is 100% compatible, that means no bugs, no glitches, not having to toggle between propietary and Xorg because it decided to crapout on me. Thanks for your help

    Read the article

  • Dual monitor not working after an update

    - by Nimonika
    I did a package manager update yesterday and it turns out that my dual monitor setup has stopped working. I have poor vision so I really need to connect to a much bigger screen, but since yesterday, when I connect the screen to my laptop, the screen does not automatically reset itself to the laptop display. Even after lots of trial and error with the display settings, I am getting different dispalys on the laptop and external screen and right now only the big screen is active while the laptop has blanked out. Please can someone help me setup my dual screens for 11.10 properly. lspci -v | grep -i vga output 00:02.0 VGA compatible controller: Intel Corporation Mobile 4 Series Chipset Integrated Graphics Controller (rev 07) (prog-if 00 [VGA controller]

    Read the article

  • poor font rendering in browsers on 12.04

    - by Alecs
    I'm using a Lenovo B570e. I've installed the mesa-utils for the graphic driver because it wasn't recognized by default. Now I have this: graphic card : Intel® Sandybridge Mobile x86/MMX/SSE2. I don't know if it's just my impression but I think all the images are a little bit pixelated and I think it have something to do with the drivers. Maybe is not the correct driver associated with it, I'm saying this because when I was on Windows 7 the name of the driver was something like " Intel hd graphics..." Any suggestions? Thank you! Later edit: This seems to happen only in browsers.

    Read the article

  • Screen gets garbled on some web sites

    - by user10565
    I have a Gateway notebook with graphics card 01:05.0 VGA compatible controller: ATI Technologies Inc RS690M [Radeon X1200 Series] with open source driver Linux version 2.6.32 -28 - generic. No other operating system on the computer. When I am using firefox to browse the web, everything normally works just fine except that when I attempt to access some particular web pages the screen completely messes up going mostly white with various streaks, etc., although I can access other pages of the same site without problems. When I run the cursor over the garbled screen, bits of the image recompose themselves, at least partially, and I can continue to open the applications window, or turn the computer off, or open the terminal, or take screen shots, etc., although all menus are unreadable. Also, when I zoom in on Google Earth the screen completely messes up. At all other times, there are no apparent problems. Any ideas?

    Read the article

  • Torchlight is missing some textures

    - by AKAGoodGravy
    I am using a Radon HD6850 and just bought the indi bundle. Torchlight is missing all face textures (and face textures only). I have had problems with shaders before (graphic would be displayed as black), but what I am seeing now is a simple removal of the texture, I can see through the head. The fps is fine on full graphics so it doesn't seem to be too much of a driver issue (I am using the drivers that came with ubuntu). I downloaded the game through the USC. Has anyone had this issue and managed to fix it? Running Ubuntu 12.04 LTS

    Read the article

  • adjust the brightness in 13.04

    - by dinesh2707
    I have a dual graphics card: intel 4000 and nvidia 630m... I installed bumblebee and intel drivers from their website. I thought the brightness problem will be solved by installing these drivers. But, brightness works only for two steps: that is, dim, medium, high... But, it doesn't increase step wise. Can you please help me sort this problem? I'm a newbie to ubuntu and this is my first trial. So, please give me a detailed guide... Thank you :) EDIT: My model is HP DV6 7040tx. My Laptop Model

    Read the article

  • Graphics and USB devices freezing soon after OS loads

    - by Andrew
    I run Ubuntu/Windows dual boot. Last night I started the upgrade to Ubuntu 12.04, and my computer has not worked since in either Windows or Ubuntu. Here's what I got when I rebooted after the upgrade, and continue to get every time I boot: Gets to GRUB screen OK. Choose Ubuntu - black screen or crazy purple lines. At first I assumed something went wrong with the upgrade (often happens). Choose Windows - works fine, I log in, but soon after that the graphics freeze (sometimes with purple artifacts). The keyboard and mouse (both USB) also lose power at the same instant, and none of the USB ports have power to them. This happens sooner or later every time I boot. Update: the HDD also appears to lose power at the same point. I have tried a live CD, but my computer refuses to boot any CD even after disabling all other boot options in the BIOS. I have disconnected everything except keyboard, mouse, graphics card with one monitor, one RAM sick and HDD; no change. I also took the little battery out to reset CMOS. I am pretty sure no matter how wrong the Ubuntu upgrade went, it wouldn't cause the above symptoms in Windows. So the only explanation I can think of is that a hardware failure occurred at the same time. Some possible causes of this I can think of are: A couple of days before this, I added a third screen (which worked fine). About a week before, my house lost power in a storm (no ill effects over the past few days though). What can I do, other than buy a new motherboard/CPU and hope it works? Unfortunately I don't have another box to swap parts into to test at the moment.

    Read the article

  • Best practise for Progress Bar in Python's PyGTK

    - by Matthew Walker
    I would like to get feedback on others' opinions of best practice for how to implement a progress bar in Python's PyGTK. The work that the progress bar was to represent was very significant computationally. Therefore, I wanted the work to be done in a separate process (thus giving the operating system the possibility to run it on a different core). I wanted to be able to start the work, and then continue to use the GUI for other tasks while waiting for the results. I have seen many people asking this question indirectly, but I have not seen any concrete expert advice. I hope that by asking this question we will see a community's combined expertise. I have read the FAQ and I hope that "Programmers" is the right place to ask.

    Read the article

  • My ubuntu runs in slow motion on ubuntu 13.04 asus 1015cx

    - by thj
    My specs: my Asus EeePC 1015CX certified by ubuntu 12.04 lts asus 1015 cx with gma 3600 graphic card intel atom dual core ubuntu 13.04 My problem: I need to use ubuntu 13.04, not 12.04. ubuntu 13.04 runs in slow motion on default unity desktop, is it graphic card problem? when i clicked file, runs in very slow animation on unity. windows comes up in very slow motion I can't even use ubuntu 13.04 because so slow. what's minimal requirement of ubuntu 13.04 or unity desktop? why is it so slow? what kind of processes should i take to fix it? I tried by myself but no good result. Need some help anyway. Any ideas ? maybe i should change hardware lol? my Asus EeePC 1015CX certified by ubuntu 12.04 lts !

    Read the article

  • How do I generate a core dump from apache

    - by blockhead
    I have a script which intermittently returns a white screen of death in firefox and Error 324 (net::ERR_EMPTY_RESPONSE): Unknown error. chrome. When I try to access the script using a PHP HTTP client (like Zend_Http_Client), intermittently I get an exception (sorry I don't have the exact message on me at the moment). I suspect a segfault. This is further buttressed by the lines in my error log that look like this: [Thu Mar 18 16:03:02 2010] [notice] child pid 845 exit signal Segmentation fault (11) Now, I'm running RedHat, and I know that RedHat doesn't generate core dumps out-of-the-box. I followed the instructions here http://kbase.redhat.com/faq/docs/DOC-5353, but I'm not seeing any core dumps. How do I generate a core dump?

    Read the article

  • upgraded "Compiz' and "Unity", now no Unity 3D on screen

    - by user18432
    Today when I logged in to my Ubuntu 12.04, the update manager told me of some upgrades. Compiz and Unity were in those upgrades. After I installed the upgrades, I can no longer get the Unity panel on the left side of screen or the systray at the top of screen. I now have to run Ubuntu 12.04 with Unity 2D. My laptop is a HP Pavilion dv9000 with Nvidia GeForce Go 7600 video. I tried to run "unity --reset" but it says there are serious issues with compiz. I have cut & pasted the read out from the terminal below. [09:35:02] xxxxxxx@L01U1204:~$ unity --reset unity-panel-service: no process found Checking if settings need to be migrated ...no Checking if internal files need to be migrated ...no Backend : gconf Integration : true Profile : unity Adding plugins Initializing core options...done compiz (core) - Warn: failed to receive ConfigureNotify event on 0x2e00004 compiz (core) - Warn: failed to receive ConfigureNotify event on 0x580005a compiz (core) - Warn: failed to receive ConfigureNotify event on 0x3600006 compiz (core) - Warn: failed to receive ConfigureNotify event on 0x3200255 compiz (core) - Warn: failed to receive ConfigureNotify event on 0x1600002 compiz (core) - Warn: failed to receive ConfigureNotify event on 0x1400002 Initializing composite options...done Initializing opengl options...done Initializing decor options...done Initializing vpswitch options...done Initializing snap options...done Initializing mousepoll options...done Initializing resize options...done Initializing place options...done Initializing move options...done Initializing wall options...done Initializing grid options...done I/O warning : failed to load external entity "/home/brwright/.compiz/session/10afaca1703486b216133648409481824100000130110002" Initializing session options...done Initializing gnomecompat options...done Initializing animation options...done Initializing fade options...done Initializing unitymtgrabhandles options...done Initializing workarounds options...done Initializing scale options...done compiz (expo) - Warn: failed to bind image to texture Initializing expo options...done Initializing ezoom options...done compiz (core) - Error: Couldn't load plugin '/usr/lib/compiz/libunityshell.so' : /usr/lib/compiz/libunityshell.so: undefined symbol: _ZNK5unity4dash10Controller6windowEv compiz (core) - Error: Couldn't load plugin 'unityshell' compiz (core) - Warn: unhandled ConfigureNotify on 0x7000090! compiz (core) - Warn: this should never happen. you should probably file a bug about this. compiz (core) - Warn: unhandled ConfigureNotify on 0x7000093! compiz (core) - Warn: this should never happen. you should probably file a bug about this. compiz (core) - Warn: unhandled ConfigureNotify on 0x7000096! compiz (core) - Warn: this should never happen. you should probably file a bug about this. compiz (core) - Warn: unhandled ConfigureNotify on 0x7000099! compiz (core) - Warn: this should never happen. you should probably file a bug about this. compiz (core) - Warn: unhandled ConfigureNotify on 0x700009c! compiz (core) - Warn: this should never happen. you should probably file a bug about this. compiz (core) - Warn: unhandled ConfigureNotify on 0x700009f! compiz (core) - Warn: this should never happen. you should probably file a bug about this. Initializing annotate options...done Initializing blur options...done Initializing clone options...done Initializing colorfilter options...done Initializing commands options...done Initializing cube options...done Initializing imgjpeg options...done Initializing kdecompat options...done Initializing mag options...done Initializing neg options...done Initializing obs options...done Initializing opacify options...done Initializing put options...done Initializing resizeinfo options...done Initializing ring options...done Initializing rotate options...done Initializing scaleaddon options...done Initializing screenshot options...done Initializing shift options...done Initializing staticswitcher options...done Initializing switcher options...done Initializing thumbnail options...done Initializing unityshell options...done Initializing water options...done Initializing winrules options...done Initializing wobbly options...done Setting Update "main_menu_key" Setting Update "run_key" Starting gtk-window-decorator As you can see the terminal never comes back to the CI prompt. I must do a control C to get to the CI prompt, but then the OS is frozen. I have to reboot and run Unity 2D in able to do anything on my laptop. I hope I have explained this enough and provided some useful info. I am at a loss to understand what the problem is, or what exactly what is causing the problem. Is it Unity or Compiz? Can anyone help?

    Read the article

  • How open is open core and is that open enough

    <b>ZDNet:</b> "The idea is that you make the center of your product open source, but put the rest under a paid license. This is supposed to make your venture capital backers happy. You gain the benefits of open source but customers aren&#8217;t &#8220;stealing&#8221; the software."

    Read the article

  • DundeeWealth Selects Oracle CRM On Demand as Core Platform

    - by andrea.mulder
    "Oracle CRM On Demand enhances our existing Oracle platform, providing an integrated solution with incredible flexibility, mobility, agility and lowered total cost of ownership," said To Anh Tran, Senior Vice President of Business Transformation and Technology at DundeeWealth Inc. "Using Oracle as a partner in the expansion of DundeeWealth's CRM processes reinforces our client-centric approach to customer service and we believe it gives us a competitive advantage. As we begin our deployment, we are confident that Oracle is with us every step of the way." Click here to read more about more about DundeeWealth's plans.

    Read the article

  • BizTalk ESB Toolkit: Core Components and Examples

    - by Rajesh Charagandla
    The BizTalk ESB Toolkit 2.0 provides a stable and powerful platform for services that can change as fast as your business needs. The main purpose of an enterprise service bus (ESB) to is to provide a common mediation layer (the “bus”) through which all services connect. By doing so, not only can many of the problems of point-to-point service connectivity be resolved, but a new level of agile service delivery can be achieved. Author: Jon Flanders This Document can be download from here.

    Read the article

  • Need help generating a core dump from apache segfault

    - by blockhead
    I have a script which intermittently returns a white screen of death in firefox and Error 324 (net::ERR_EMPTY_RESPONSE): Unknown error. chrome. When I try to access the script using a PHP HTTP client (like Zend_Http_Client), intermittently I get an exception (sorry I don't have the exact message on me at the moment). I suspect a segfault. This is further buttressed by the lines in my error log that look like this: [Thu Mar 18 16:03:02 2010] [notice] child pid 845 exit signal Segmentation fault (11) Now, I'm running RedHat, and I know that RedHat doesn't generate core dumps out-of-the-box. I followed the instructions here http://kbase.redhat.com/faq/docs/DOC-5353, but I'm not seeing any core dumps. How do I generate a core dump?

    Read the article

  • SQL SERVER – Core Concepts – Elasticity, Scalability and ACID Properties – Exploring NuoDB an Elastically Scalable Database System

    - by pinaldave
    I have been recently exploring Elasticity and Scalability attributes of databases. You can see that in my earlier blog posts about NuoDB where I wanted to look at Elasticity and Scalability concepts. The concepts are very interesting, and intriguing as well. I have discussed these concepts with my friend Joyti M and together we have come up with this interesting read. The goal of this article is to answer following simple questions What is Elasticity? What is Scalability? How ACID properties vary from NOSQL Concepts? What are the prevailing problems in the current database system architectures? Why is NuoDB  an innovative and welcome change in database paradigm? Elasticity This word’s original form is used in many different ways and honestly it does do a decent job in holding things together over the years as a person grows and contracts. Within the tech world, and specifically related to software systems (database, application servers), it has come to mean a few things - allow stretching of resources without reaching the breaking point (on demand). What are resources in this context? Resources are the usual suspects – RAM/CPU/IO/Bandwidth in the form of a container (a process or bunch of processes combined as modules). When it is about increasing resources the simplest idea which comes to mind is the addition of another container. Another container means adding a brand new physical node. When it is about adding a new node there are two questions which comes to mind. 1) Can we add another node to our software system? 2) If yes, does adding new node cause downtime for the system? Let us assume we have added new node, let us see what the new needs of the system are when a new node is added. Balancing incoming requests to multiple nodes Synchronization of a shared state across multiple nodes Identification of “downstate” and resolution action to bring it to “upstate” Well, adding a new node has its advantages as well. Here are few of the positive points Throughput can increase nearly horizontally across the node throughout the system Response times of application will increase as in-between layer interactions will be improved Now, Let us put the above concepts in the perspective of a Database. When we mention the term “running out of resources” or “application is bound to resources” the resources can be CPU, Memory or Bandwidth. The regular approach to “gain scalability” in the database is to look around for bottlenecks and increase the bottlenecked resource. When we have memory as a bottleneck we look at the data buffers, locks, query plans or indexes. After a point even this is not enough as there needs to be an efficient way of managing such large workload on a “single machine” across memory and CPU bound (right kind of scheduling)  workload. We next move on to either read/write separation of the workload or functionality-based sharing so that we still have control of the individual. But this requires lots of planning and change in client systems in terms of knowing where to go/update/read and for reporting applications to “aggregate the data” in an intelligent way. What we ideally need is an intelligent layer which allows us to do these things without us getting into managing, monitoring and distributing the workload. Scalability In the context of database/applications, scalability means three main things Ability to handle normal loads without pressure E.g. X users at the Y utilization of resources (CPU, Memory, Bandwidth) on the Z kind of hardware (4 processor, 32 GB machine with 15000 RPM SATA drives and 1 GHz Network switch) with T throughput Ability to scale up to expected peak load which is greater than normal load with acceptable response times Ability to provide acceptable response times across the system E.g. Response time in S milliseconds (or agreed upon unit of measure) – 90% of the time The Issue – Need of Scale In normal cases one can plan for the load testing to test out normal, peak, and stress scenarios to ensure specific hardware meets the needs. With help from Hardware and Software partners and best practices, bottlenecks can be identified and requisite resources added to the system. Unfortunately this vertical scale is expensive and difficult to achieve and most of the operational people need the ability to scale horizontally. This helps in getting better throughput as there are physical limits in terms of adding resources (Memory, CPU, Bandwidth and Storage) indefinitely. Today we have different options to achieve scalability: Read & Write Separation The idea here is to do actual writes to one store and configure slaves receiving the latest data with acceptable delays. Slaves can be used for balancing out reads. We can also explore functional separation or sharing as well. We can separate data operations by a specific identifier (e.g. region, year, month) and consolidate it for reporting purposes. For functional separation the major disadvantage is when schema changes or workload pattern changes. As the requirement grows one still needs to deal with scale need in manual ways by providing an abstraction in the middle tier code. Using NOSQL solutions The idea is to flatten out the structures in general to keep all values which are retrieved together at the same store and provide flexible schema. The issue with the stores is that they are compromising on mostly consistency (no ACID guarantees) and one has to use NON-SQL dialect to work with the store. The other major issue is about education with NOSQL solutions. Would one really want to make these compromises on the ability to connect and retrieve in simple SQL manner and learn other skill sets? Or for that matter give up on ACID guarantee and start dealing with consistency issues? Hybrid Deployment – Mac, Linux, Cloud, and Windows One of the challenges today that we see across On-premise vs Cloud infrastructure is a difference in abilities. Take for example SQL Azure – it is wonderful in its concepts of throttling (as it is shared deployment) of resources and ability to scale using federation. However, the same abilities are not available on premise. This is not a mistake, mind you – but a compromise of the sweet spot of workloads, customer requirements and operational SLAs which can be supported by the team. In today’s world it is imperative that databases are available across operating systems – which are a commodity and used by developers of all hues. An Ideal Database Ability List A system which allows a linear scale of the system (increase in throughput with reasonable response time) with the addition of resources A system which does not compromise on the ACID guarantees and require developers to learn new paradigms A system which does not force fit a new way interacting with database by learning Non-SQL dialect A system which does not force fit its mechanisms for providing availability across its various modules. Well NuoDB is the first database which has all of the above abilities and much more. In future articles I will cover my hands-on experience with it. Reference: Pinal Dave (http://blog.SQLAuthority.com) Filed under: PostADay, SQL, SQL Authority, SQL Query, SQL Server, SQL Tips and Tricks, T SQL, Technology Tagged: NuoDB

    Read the article

  • Server Systems for SQL Server 2012 per core licensing

    - by jchang
    Until recently, the SQL Server Enterprise Edition per processor (socket) licensing model resulted in only 2 or 3 server system configurations being the preferred choice. Determine the number of sockets: 2, 4 or 8. Then select the processor with the most compute capability at that socket count level. Finally, fill the DIMM sockets with the largest capacity ECC memory module at reasonable cost per GB. Currently this is the 16GB DIMM with a price of $365 on the Dell website, and $240 from Crucial. The...(read more)

    Read the article

< Previous Page | 50 51 52 53 54 55 56 57 58 59 60 61  | Next Page >