Search Results

Search found 6029 results on 242 pages for 'discrete graphics'.

Page 21/242 | < Previous Page | 17 18 19 20 21 22 23 24 25 26 27 28  | Next Page >

  • How can I replicate the look and limitations of the Super NES?

    - by Mikalichov
    I am looking to produce graphics with the same limitations / look that in the Super Nes era. I am specifically looking for graphics similar to Chrono Trigger / FF6. It would be a lot easier to do if I had an idea of the resolution / dpi I am supposed to use. I found that the technical specs for the SNES are: Progressive: 256 × 224, 512 × 224, 256 × 239, 512 × 239 Interlaced: 512 × 448, 512 × 478 But even by using these resolutions, it is pointless if I set it at 72dpi, as I will still have possibly very detailed graphics (that is the main thing, I don't want detailed graphics, I want to go pixelated). I figured it might be related to the sprite size limit, i.e.: Sprites can be 8 × 8, 16 × 16, 32 × 32, or 64 × 64 pixels, each using one of eight 16-color palettes and tiles from one of two blocks of 256 in VRAM. Up to 32 sprites and 34 8 × 8 sprite tiles may appear on any one line. This would work for sprites (characters, objects), but what about maps? Are they built entirely from 8x8 tiles? And then, at what resolution is the end result displayed? It might seem like I am giving the question and answers at the same time, but all of these are suppositions I am making, so could someone confirm or correct them?

    Read the article

  • Kubuntu never gets to the login screen

    - by searchfgold6789
    I am not sure if Xorg is using the right device... The system in question is an A4-1400 CPU with built-in Radeon Graphics. I added a PCIe Radeon HD 5750, and changed the BIOS so it would use that card automatically. Now it appears that I can see output from the TV plugged into the HDMI port, and boot into recovery mode, but cannot get to a desktop - X doesn't start. It just freezes at the Kubuntu splash screen. I have no idea how to get X to use the proper graphics card. I pasted the Xorg.log to http://paste.ubuntu.com/6261137/. And the Xorg.conf generated by aticonfig --initial is at http://paste.ubuntu.com/6261161/ Also, see lspci and lshw -c display outputs. I can provide more information, but something does tell me X is not using the right graphics card and I've no idea where to begin how to fix it... I faced a similar issue before with another distribution but found no information anywhere on how to make X use a specific graphics card. Or maybe it's not that simple.

    Read the article

  • Day 5 - Tada! My Game Menu Screen Graphics

    - by dapostolov
    So, tonight I took some time to mash up some graphics for my game menu screen. My artistic talent sucks...but here goes nothing...voila, my menu screen!! The Menu Screen The screen above is displaying 4 sprites, even though it looks like maybe 7... I guess one of the first things for me to test in the future is ... is it more memory efficient (and better frame rate) to draw one big background image OR tp paint the screen black, and place each sprite in set locations? To display the 4 sprites above, I borrowed my code from yesterday ... I know, tacky, but...I wanted to see it, feel it. Do you feel it? FEEL IT! (homer voice & shakes fist) Note: the menu items won't scale properly as it stands with this code, well pretty much they do nothing except look pretty... Paint.Net & Google Fun So how did I create that image above? Well, to create the background and 3 menu items I used Paint.Net. Basically, I scoured Google images for: a stone doorway, a stone pillar, an old book, a wizards hat, and...that's it pretty much it! I'll let you type in those searches and see if you can locate the images I used. I know, bad developer...but I figured since I modified the images considerably it doesn't count...well for a personal project it shouldn't count...*shrug* Anyhow, I extracted each key assest I wanted from each image and applied lots of matting, blurring, color changes, glow effects and such. Then, using my vivid imagination I placed / composed each of the layered assets into the mashed up the "scene" above. Pretty cool, eh? Hey, did you know, the cool mist effect is actually a fire rendition in Paint.net? I set it to black & white with opacity set next to nothing. I'm also very proud of the yellow "light" in the stone doorway. I drew that in and then applied gausian blur to it to give it the effect of light creeping out around the door and into the room...heheh. So did I achieve the dark, mysterious ritual as I stated in my design doc? I think I had a great stab at it! Maybe down the road I can get a real artist to crank out some quality graphics for the game... =) So, What's Next? Well, I don't have that animated brazier yet...however, I thought it would be even cooler if I can get that door pulsing that yellow light and it would be extremely cool to have the smoke / mist moving across the screen! Make the creative ideas stop!! (clutches head) haha! I'm having great fun working on this project =) I recommend others giving something like this a try, it's really fulfilling. OK. Tomorrow... I think I'm going to start creating some game / menu objects as per the design doc, maybe even get a custom mouse cursor up on the screen and handle a couple of mouse events, and lastly, maybe a feature to toggle a framerate display... D.

    Read the article

  • Rendering of graphics different depending on DisplayObject position

    - by jedierikb
    When drawing vertical lines with a non-integer x-value (e.g., 1.75) to a sprite, the lines are drawn differently based on the non-integer x-value of the sprite. In the picture below are two pairs of very close together vertical lines. As you can see, they look very different. This is frustrating, especially when animating the sprite. Any ideas how ensure that sprites-with-non-integer-positions' graphics will visually display the same way regardless of the sprite position? package { import flash.display.Sprite; import flash.display.StageAlign; import flash.display.StageScaleMode; import flash.events.Event; public class tmp extends Sprite { private var _sp1:Sprite; private var _sp2:Sprite; public function tmp( ):void { stage.align = StageAlign.TOP_LEFT; stage.scaleMode = StageScaleMode.NO_SCALE; _sp1 = new Sprite( ); drawButt( _sp1 ); _sp1.x = 100; _sp1.y = 100; _sp2 = new Sprite( ); drawButt( _sp2 ); _sp2.x = 100; _sp2.y = 200; addChild( _sp1 ); addChild( _sp2 ); addEventListener( Event.ENTER_FRAME, efCb, false, 0, true ); } private function efCb( evt:Event ):void { var nx:Number = _sp2.x + .1; if (nx > 400) { nx = 100; } _sp2.x = nx; } private function drawButt( sp:Sprite ):void { sp.graphics.clear( ); sp.graphics.lineStyle( 1, 0, 1, true ); sp.graphics.moveTo( 1, 1 ); sp.graphics.lineTo( 1, 100 ); sp.graphics.lineStyle( 1, 0, 1, true ); sp.graphics.moveTo( 1.75, 1 ); sp.graphics.lineTo( 1.75, 100 ); } } }

    Read the article

  • why PaintComponent event in Java happen everytime I use its Graphics Event?

    - by Pooya
    Consider this code: public class StateChartPanel extends JPanel { private LightContext LC; public StateChartPanel(LightContext lc){ LC=lc; } public void paintComponent( Graphics G ){ super.paintComponent( G ); LC.DrawStateChart((Graphics2D)G); } } StateChartPanel is a panel to draw something (a state chart). It sends its Graphics object to LC which use it to draw shapes but whenever it draws something the PaintComponent event of StateChartPanel happens again and it causes my application to hang.

    Read the article

  • Ubuntu 12.10 - Graphics driver shows Gallium 0.4 instead of intel HD

    - by nlooije
    I have a hybrid Intel/Nvidia system using bumblebee on Ubuntu 12.10 with specs: Clevo W150HR, SB i7-2720qm, 8GB RAM, 128GB Crucial M4 SSD + 500GB HDD It is reported to use software rendering (through Gallium driver) rather than accelerated through the intel HD driver. Reinstalling the intel driver has no effect: sudo apt-get install --reinstall xserver-xorg-video-intel The result is a rather sluggish desktop. Is there any way to blacklist the gallium driver? Or force the intel driver instead? Edit 1 - Ubuntu 12.04 shows the driver as Intel SandyBridge out of the box, but 12.10 does not even after running the above command. Edit 2 - Xorg.0.log shows: (--) intel(0): Integrated Graphics Chipset: Intel(R) Sandybridge Mobile (GT1) (WW) intel(0): Disabling hardware acceleration on this pre-production hardware. Edit 3 - It turns out that SandyBridge rev.7 systems are known to be unstable per this link. Accordingly in the xserver-xorg-video-intel if it detects this rev. it disables it with the warning in above log.

    Read the article

  • Often "running in low-graphics mode” with NVidia GeForce GT 430

    - by user42228
    Often (every fifth time perhaps) when I turn on my system, I'm greeted with the "running in low-graphics mode” window. Sometimes all I have to do is reboot and the system will start correctly, sometimes there can be a longer series of failed starts. If I go to the terminal and log in, I can run startx and my desktop will appear without problems. At the end of /var/log/Xorg.1.log I see this: [ 77.459] (EE) Screen(s) found, but none have a usable configuration. [ 77.459] Fatal server error: [ 77.459] no screens found [ 77.459] Please consult the The X.Org Foundation support at http://wiki.x.org Is the NVidia creating the appropriate XOrg configuration dynamically on every boot, or what could explain that the configuration error above only occurs sometimes?

    Read the article

  • iPhone image asset recommended resolution/dpi/format

    - by Matthew
    I'm learning iPhone development and a friend will be doing the graphics/animation. I'll be using cocos2d most likely (if that matters). My friend wants to get started on the graphics, and I don't know what image resolution or dpi or formats are recommended. This probably depends on if something is a background vs. a small character. Also, I know I read something about using @2x in image file names to support high res iphone screens. Does cocos2d prefer a different way? Or is this not something to worry about at this point? What should I know before they start working on the graphics?

    Read the article

  • Rebooting problem

    - by Vinoth
    After restarting my ubuntu PC, I am getting a dialog window showing "Your system is running in low graphics mode" and after clicking on 'OK', I was prompted with three options 1. Run in low graphics mode for just one session 2. Reconfigure graphics 3. Troubleshoot the error 4. Exit to console login When I troubleshoot the error, its showing the black screen with following lines which I remember: we do not own /var/log/pm-powersave.log ....................................... couldn't write bytes: Broken pipe ....................................... ....................................... Starting System V runlevel compatibility got strucked at this point. Actually before restarting, I have changed the owner for all the files under the directory "/var/log" and "/var/lib". Is there any relation with the above issue. Please help me out to resolve the issue. Its urgent.

    Read the article

  • Split screen and other issues on Ubuntu 11.10 with ATI graphics card

    - by garus
    Ever since updating my Ubuntu 11.04 to 11.10 I experience issues with graphics. The biggest is the "split screen" effect where my desktop is shifted to the right, resulting in having Unity bar in the middle of the screen. As shown here: http://i.imgur.com/I8nmZ.jpg This changes from boot to boot, sometimes it's on the left, sometimes in the middle. What I tried: Removing fglrx drivers completely Installing post-update version of them, but installation is broken ATM in the Ubuntu, so no go. No one is even trying to fix it (bug report https://bugs.launchpad.net/ubuntu/+source/nvidia-common/+bug/873058 and a couple of duplicates out there) Also using the open source "radeon" driver results in the same (I have a better successful boot ratio with this one, proprietary rarely lets me boot) Other artifacts are: serious screen tearing weird lines flickering in random places lagginess Did anyone experience that? My specs: Ubuntu 11.10 AMD Radeon HD 6950 1GB

    Read the article

  • unmet dependencies updating Intel graphics driver

    - by Nick Jacobs
    I am trying to update my graphics driver by running: sudo apt-get install xserver-xorg-video-intel But I keep getting this message: Some packages could not be installed. This may mean that you have requested an impossible situation or if you are using the unstable distribution that some required packages have not yet been created or been moved out of Incoming. The following information may help to resolve the situation: The following packages have unmet dependencies: xserver-xorg-video-intel : Depends: xorg-video-abi-11 E: Unable to correct problems, you have held broken packages. When I run: sudo apt-get install -f Nothing happens, because the program doesn't even get half way to installing. I am running Lubuntu 12.04 on a HP Mini 110 netbook. I have tried changing to several different ppa's but that doesn't work. I want to update my driver so that my mouse stops flickering and ScummVM stops flickering every minute.

    Read the article

  • "The system is running in low graphics mode", unable to connect to the Internet

    - by Blah
    When I try to log onto Ubuntu, I get a notice that reads: The system is running in low graphics mode. I searched for a solution and there are a slew of answers. However, many of them seem to suggest reinstalling different things. After pressing Ctrl+Alt+F1 and getting to a terminal, I attempt commands like sudo apt-get install example. Upon doing so I get notices that say things such as could not resolve host which I am assuming means I'm not connected to the Internet. Since I can not connect to the Internet, none of these responses are helpful to me. What can I do?

    Read the article

  • After reinstalling ATI graphics drivers, my keyboard and mous aren't working anymore

    - by Lifelike27
    On Ubuntu 11.04 I tried to install the open source drivers for my laptops ATI Mobility Radeon 5470M but I messed that up a bit and lost xserver. Now I've managed to solve that problem with this and by downloading the ATI proprietary drivers and install those manually. Now, when Ubuntu loads up I get to the login screen but I can't use my mouse or my keyboard (usb keyboard and mouse doesn't work either). If I use the recovery console, then login with that and then run 'startx'. I can login fine (though Unity doesn't show, the graphics seem to be working because it shows the fading animation of libnotify), but I can't type or move my mouse.

    Read the article

  • Ubuntu 13.04/13.10 - low graphics mode/black screen after login

    - by Richard Stokes
    Since last week, my Ubuntu 13.04 installation started booting up in low graphics mode out of nowhere. I tried a number of solutions found on this site but to no avail. The other night, I decided to upgrade to 13.10 to see if the issue would be fixed in the process. Since then, the system starts fine, but when I login to my account, I'm just presented with a black screen. I switched to the terminal and did any necessary updates, but to no avail. I don't even know where to start debugging this issue, so I'd appreciate if someone could point me towards the first place to start looking?

    Read the article

  • aticonfig detects all graphics cards, but amdcccle shows the cards as disabled

    - by DevZer0
    I been having the strangest problem since few days ago. I have 3 ATI graphics cards and the aticonfig --list-adapter shows the following output * 0. 01:00.0 AMD Radeon HD 7900 Series 1. 02:00.0 AMD Radeon HD 7900 Series 2. 05:00.0 AMD Radeon HD 7900 Series when i do aticonfig --adapter=all --initial -f it generates the correct X config also. but after rebooting the output only show on primary monitor and when i look in amdcccle it shows the other 2 adapters as disabled. I tried with monitors attached and dummy plugs both but the situation doesn't change. Any idea whats causing this? also right clicking and making the adapter enabled in amdcccle and saving changes causes the X config to only have the adapter section and no screen section. after reboot the situation stays the same.

    Read the article

  • xrandr shows VGA1 as disconnected

    - by Felix
    I have a Thinkpad W520 with Nvidia Optimus graphics. I have disabled the Nvidia card in BIOS (by selecting "integrated graphics"), so I'm running only on the integrated Intel graphics. I get full 3D acceleration, which would suggest the drivers are properly installed. However, I'm not able to use an external monitor. With the external monitor connected and turned on, running xrandr always gives: $ xrandr Screen 0: minimum 320 x 200, current 1920 x 1080, maximum 8192 x 8192 LVDS1 connected 1920x1080+0+0 (normal left inverted right x axis y axis) 344mm x 193mm 1920x1080 60.0*+ 59.9 50.0 1680x1050 60.0 59.9 1600x1024 60.2 1400x1050 60.0 1280x1024 60.0 1440x900 59.9 1280x960 60.0 1360x768 59.8 60.0 1152x864 60.0 1024x768 60.0 800x600 60.3 56.2 640x480 59.9 VGA1 disconnected (normal left inverted right x axis y axis) What gives? It sees the VGA1 port (to which the external display is connected), but it appears disconnected. I have tried forcing a resolution as per these instructions, but when I do that X becomes unresponsive and I have to Ctrl-Alt-F1 and restart it.

    Read the article

  • Depending on another open source library: copy/paste code or include

    - by user5794
    I'm working on a large class and started implementing new features that need graphics. I started writing the graphics functions myself, but I know that open source libraries exist that can provide me with this functionality without me having to write it myself. The problem is that I prefer the class to be self-sufficient and not dependent on any other library. If I don't write it myself, I would have to ask the user to make sure a graphics library is already installed (less user-friendly). If I write it myself, I do a lot more work than I have to. I could also copy/paste some of the relevant code into my own class, but not sure about the disadvantages of doing this (it's an open source library that matches my license, so I'm not concerned with legality, just programming-wise if there are disadvantages). So what should I do: copy paste code from the external library write the code myself so it's truly self-sufficient ask the user to download and install another library

    Read the article

  • After reinstalling ATI graphics drivers, my keyboard and mouse aren't working anymore

    - by Lifelike27
    On Ubuntu 11.04 I tried to install the open source drivers for my laptops ATI Mobility Radeon 5470M but I messed that up a bit and lost xserver. Now I've managed to solve that problem with this and by downloading the ATI proprietary drivers and install those manually. Now, when Ubuntu loads up I get to the login screen but I can't use my mouse or my keyboard (usb keyboard and mouse doesn't work either). If I use the recovery console, then login with that and then run 'startx'. I can login fine (though Unity doesn't show, the graphics seem to be working because it shows the fading animation of libnotify), but I can't type or move my mouse.

    Read the article

  • Bumblebee optirun appears to depend on Intel

    - by user206398
    I have a Lenovo T420 with Intel and Nvidia graphics. On upgrade to Ubuntu Saucy, I had to purge and reinstall bumblebee-nvidia to get beyond optirun failing to find a GPU driver. Now, "optirun glxgears" and "optirun sol" succeed, but optirun fails on 2 Virtual Life viewers that it supported in the past, Cool VL (CoolVLViewer-1.26.8.34-Linux-x86) and Imprudence (Imprudence 1.4.0 beta2). In both cases, the error output is huge, but it starts with libGL error: failed to load driver: i965 and libGL error: failed to load driver: swrast From the little I can discover, i965 is an Intel graphics driver, which should not be invoked at all. I haven't found any information about swrast. I suspect that some of the X configuration associated with Bumblebee has some Intel dependence that is invoked on certain library calls, but not others. I haven't discovered any definite information on this line. The Cool VL Viewer runs without optirun, but complains about the insufficiency of the Intel graphics.

    Read the article

  • The system is running in low graphics mode error before ubuntu gets installed

    - by hellodear
    I am using Bootable USB for installing Ubuntu with Windows 8.1. I have put pendrive and then booting my USB. When it shows me to try Ubuntu or install Ubuntu. Then I press Try ubuntu and then it says" "The system is running in low graphics mode". Then I press Ok. Then it shows 4 options. Then again I click Ok. Then it shows a black screen and nothing happens. I have tried all possible answers provided in AU. What should I do? Please help. I am using Windows 8.1 with dedicated graphic card (Video card - AMD Radeon HD 8670M). Installing 12.04 LTS with pre-installe Windows 8.1 in a Dell Laptop 3537 inspiron.

    Read the article

  • lightdm.conf content erased, now stuck in low graphics mode

    - by user79318
    This evening I was attempting to disable the guest account and something went awry. Currently on boot Ubuntu enters low graphics mode. No specific error report. What did I do? Before this error occurred I added a line of code in lightdm.conf to disable the guest account. I think I may have accidentally erased the contents of lightdm.conf. Not entirely sure. I troubleshooted for the past hour using various suggestions from other Questions to no avail.

    Read the article

  • Hyper-V Blue Screens with Nvidia GeForce 8400 GS Graphics Card

    - by Mahmoud Saleh
    I am using Windows Server 2008 R2 Enterprise x64. After installing the Hyper-V role and restarting the machine, I get a blue screen error and an immediate reboot. I have Googled the issue and tracked it down to the graphics card, so I uninstalled it, and then Windows loads fine. However, after installing the graphics driver again, the Blue Screen returns. The graphics card is an Nvidia GeForce 8400 GS. Does anyone know how I can resolve this issue?

    Read the article

  • help with matlab and Discrete Fourier transform

    - by user504363
    Hi all I have previous experience with Matlab, but the problem that I face some problems in apply a problem in (DSP : Digital signal processing) which is not my study field, but I must finish that problems in days to complete my project. all i want is help me with method and steps of solving this problem in matlab and then I can write the code with myself. the problem is about the signal x(t) = exp(-a*t); 1) what's the Discrete Fourier transform of the sampled signal with sample rate fs 2) if a=1 and fs =1 , plot the amplitude spectrum of sampled signal 3) fix the sampling frequency at fs = 1(hz) [what's it mean ?] and plot the magnitude of the Fourier Transform of the sampled signal at various values of a thanks

    Read the article

  • Discrete event simulation framework for .NET

    - by Kuba
    Does anyone have an experience with some discrete event simulation library that could be used in .NET (C#)? Despite the basic functionality for queing events and dispatching them, it would be fine to have some non-deterministic behavior (e.g. failures simulation). I have some tips and I am even considering to write my own, but first, I would like to collect some recomendations. Thanks. Additional info: i'm not looking explicitly for free product, however, the prize matters :) Just to precise the field i need to map, here is the example of a product: http://www.holushko.com/index.html

    Read the article

  • How do I keep from running out of memory on graphics for an Android app?

    - by user279112
    I've been working on an Android app in Eclipse, and so far, my program hasn't really grown past midget size. However I've already run into an issue with an Out of Memory error. You see, I've been using graphics comprised solely of bitmaps and PNGs in this program, and recently, when I tried to add a little bit more functionality to the program (mainly including a few more bitmaps and causing an extra sprite to be created), it started crashing in the graphics thread's constructor - sprite's constructor. When I tracked the problem down, it turned out to be an Out of Memory error that is seemingly caused by adding too many picture files to the program and creating Drawables out of them. This would be a problem, as I really don't have that many picture resources worked into that program...maybe 20 or so. I haven't even started to include sound yet. These images aren't all that fancy. My questions are this: 1) Are programs for the Android phone really that limited on how much memory they can employ, or is it probably something other than the 20-30 resource pictures causing that error? 2) If the memory for Android apps is so awful it can't even handle 20-30 picture resources being loaded into Drawables that exist at the same time, then how in the world are you supposed to make decent graphics and sound for that thing? Thanks.

    Read the article

< Previous Page | 17 18 19 20 21 22 23 24 25 26 27 28  | Next Page >