Search Results

Search found 1202 results on 49 pages for 'specs'.

Page 13/49 | < Previous Page | 9 10 11 12 13 14 15 16 17 18 19 20  | Next Page >

  • Why does my Ubuntu freeze in the middle of something with a blank screen?

    - by mohamed
    I am facing a serious problem as my Dell Optiplex 745 loaded with Ubuntu 12.04 32bit freezes with a blank screen no mouse cursor no keyboard activity in many occasions. Sometimes when I open firefox, or software center, I really can't tell what the main problem is, but what i am certain about is it i have to do a hard reset to reboot. I tried to use a live-cd to boot but it did the same, tried the recovery console but couldn't find a way to fix from there (I am a new user) suspected a hardware problem but windows runs flawlessly so I believe its something related to ubuntu. Specs: Intel Celeron D 3.06 GHZ 3.0 GB ram Intel built in graphics Q965 1 TB HDD Bios Phoenex 2.6.4 Please help I really love ubuntu and I don't want to go back to windows!

    Read the article

  • Does md2 support skeletal meshes?

    - by jsvcycling
    I'm creating an FPS game. I'm writing my own game engine. So far all the backend stuff is going great. I'd like to support md2 as the native file format for 3D Objects, but I also want to use skeletal meshes. Does anyone know if the md2 file format supports skeletal meshes? In-case you need to know, I'm going to use blender as my Mesh creation tool and C++ as my programming language... Thanks For got to mention, the engine is based on OpenGL... Alright, for anyone who is reading this, I just found the Doom 3 md5 specifications (http://tfc.duke.free.fr/coding/md5-specs-en.html). It gives you some help on writing a parser (see bottom of link), but the example doesn't support lighting and texture mapping (the second set of example code allows for animation). Thanks @Neverender for answering my question...

    Read the article

  • Can an agile shop every really score 12 on the Joel Test? [closed]

    - by Simon
    Possible Duplicate: Can an agile shop every really score 12 on the Joel Test? I really like the Joel test, use it myself, and encourage my staff and interviewees to consider it carefully. However I don't think I can ever score more than 9 because a few points seem to contradict the Agile Manifesto, XP and TDD, which are the bedrocks of my world. Specifically the questions about schedule, specs, testers and quiet working conditions run counter to what we are trying to create and the values that we have adopted in being genuinely agile. So my question is whether it is possible for a true Agile shop to score 12? [Note: I had this question closed on meta so I have re-posted here]

    Read the article

  • Cannot Create a connection to Data Source VB 2010 [closed]

    - by CLO_471
    I seem to be having some issues with my Visual Basic 2010. I am trying to create a connection to a data source and it is just not working. Even my old connections in my other projects are not working. When I get into VB I try and create a connection by clicking Add New Data Source Database DataSet New Connection and when I click on New Connection the screen disappears and I am not able to select anything. Does anyone know of a glitch or something? I have checked my ODBC connections and all is good and I have been able to play around with my Access connections (which I am trying to connect) and Queries and everything seems to be working fine. I have rebooted several times, uninstalled and resinstalled VB and have also repaired the entire application. I am not sure what else to try or what else to do. Any help would be much appreciated. My computer specs are XP SP3, Core2 Duo at 2.80 and 3GB RAM

    Read the article

  • Can an agile shop every really score 12 on the Joel Test?

    - by Simon
    I really like the Joel test, use it myself, and encourage my staff and interviewees to consider it carefully. However I don't think I can ever score more than 9 because a few points seem to contradict the Agile Manifesto, XP and TDD, which are the bedrocks of my world. Specifically the questions about schedule, specs, testers and quiet working conditions run counter to what we are trying to create and the values that we have adopted in being genuinely agile. So my question is whether it is possible for a true Agile shop to score 12?

    Read the article

  • How do I play a Blu Ray disc?

    - by hugocreal
    Hello All, I want play a Blu Ray Video from my hard drive with Boxee, but it chokes all the time, i try play with others video players but is the same. I think that is a 10gb .mkv file with 10Gb. Stuttering video with VLC , mplayer, and the default video player on ubuntu... I read in many Forums just can´t put this to work. Any idea? thanks. Ubuntu 10.10, My PC specs: Single Core 2Ghz ATI HD 4350 (i have installed the drivers from "Hardware Drivers"), 2G Memory

    Read the article

  • How do I play HD video without stuttering?

    - by hugocreal
    Hello All, I want play a Blu Ray Video from my hard drive with Boxee, but it chokes all the time, i try play with others video players but is the same. I think that is a 10gb .mkv file with 10Gb. Stuttering video with VLC , mplayer, and the default video player on ubuntu... I read in many Forums just can´t put this to work. Any idea? thanks. Ubuntu 10.10, My PC specs: Single Core 2Ghz ATI HD 4350 (i have installed the drivers from "Hardware Drivers"), 2G Memory

    Read the article

  • Ubuntu/Xubuntu install freeze

    - by Daniel
    I've been trying to install (dual boot) Ubuntu on my HP PC, Link to my specs (click here) I've tried to install Ubuntu with both USB and DVD, same result. The installer will launch, it will load, but then all will freeze in the middle of the loading. Well, I thought it was something wrong with Ubuntu, so I tried Xubuntu instead. Same thing, the installer will go to the loading screen, but then, it'll just freeze at the point where the "loading cursor" appears. Any help would be very much appreciated!

    Read the article

  • How do I get an Acer Aspire One D270 working?

    - by ronnie
    I was planning to buy Acer Aspire One D270 within a few days and as everybody installs linux on their netbook I was also planning to do that. Now, my question is how is Acer's hardware compatibility with linux and specifically in respect to the new Acer Aspire One D270. Has anybody tried installing linux on these new netbooks. It will be a great help if a D270 user can share his/her experience with linux usage. I read on some forums that there is some linux driver issue with Intel GMA 3600 and that people are not able to adjust their brightness. So, as I am a linux noob is this a major issue or not. Specs: RAM : 2Gb DDR3 Processor: Intel N2600(Cedar Trail) Graphics: Intel GMA 3600 HardDisk: 320Gb 5400 rpm

    Read the article

  • Design Pattern for Complex Data Modeling

    - by Aaron Hayman
    I'm developing a program that has a SQL database as a backing store. As a very broad description, the program itself allows a user to generate records in any number of user-defined tables and make connections between them. As for specs: Any record generated must be able to be connected to any other record in any other user table (excluding itself...the record, not the table). These "connections" are directional, and the list of connections a record has is user ordered. Moreover, a record must "know" of connections made from it to others as well as connections made to it from others. The connections are kind of the point of this program, so there is a strong possibility that the number of connections made is very high, especially if the user is using the software as intended. A record's field can also include aggregate information from it's connections (like obtaining average, sum, etc) that must be updated on change from another record it's connected to. To conserve memory, only relevant information must be loaded at any one time (can't load the entire database in memory at load and go from there). I cannot assume the backing store is local. Right now it is, but eventually this program will include syncing to a remote db. Neither the user tables, connections or records are known at design time as they are user generated. I've spent a lot of time trying to figure out how to design the backing store and the object model to best fit these specs. In my first design attempt on this, I had one object managing all a table's records and connections. I attempted this first because it kept the memory footprint smaller (records and connections were simple dicts), but maintaining aggregate and link information between tables became....onerous (ie...a huge spaghettified mess). Tracing dependencies using this method almost became impossible. Instead, I've settled on a distributed graph model where each record and connection is 'aware' of what's around it by managing it own data and connections to other records. Doing this increases my memory footprint but also let me create a faulting system so connections/records aren't loaded into memory until they're needed. It's also much easier to code: trace dependencies, eliminate cycling recursive updates, etc. My biggest problem is storing/loading the connections. I'm not happy with any of my current solutions/ideas so I wanted to ask and see if anybody else has any ideas of how this should be structured. Connections are fairly simple. They contain: fromRecordID, fromTableID, fromRecordOrder, toRecordID, toTableID, toRecordOrder. Here's what I've come up with so far: Store all the connections in one big table. If I do this, either I load all connections at once (one big db call) or make a call every time a user table is loaded. The big issue here: the size of the connections table has the potential to be huge, and I'm afraid it would slow things down. Store in separate tables all the outgoing connections for each user table. This is probably the worst idea I've had. Now my connections are 'spread out' over multiple tables (one for each user table), which means I have to make a separate DB called to each table (or make a huge join) just to find all the incoming connections for a particular user table. I've avoided making "one big ass table", but I'm not sure the cost is worth it. Store in separate tables all outgoing AND incoming connections for each user table (using a flag to distinguish between incoming vs outgoing). This is the idea I'm leaning towards, but it will essentially double the total DB storage for all the connections (as each connection will be stored in two tables). It also means I have to make sure connection information is kept in sync in both places. This is obviously not ideal but it does mean that when I load a user table, I only need to load one 'connection' table and have all the information I need. This also presents a separate problem, that of connection object creation. Since each user table has a list of all connections, there are two opportunities for a connection object to be made. However, connections objects (designed to facilitate communication between records) should only be created once. This means I'll have to devise a common caching/factory object to make sure only one connection object is made per connection. Does anybody have any ideas of a better way to do this? Once I've committed to a particular design pattern I'm pretty much stuck with it, so I want to make sure I've come up with the best one possible.

    Read the article

  • New iPad vs. iPad 2–Side by side comparison of hardware specification [Infographic]

    - by Gopinath
    Apple released the 3rd generation of iPad on March 7th with spectacular hardware and software specs. The new iPad is the most advanced tablet available in the market with not much of competition. The closest competitor to the new iPad is not from Android or RIM or Amazon as they are no where close to the standards of the new iPad . But the competitor is none other than previous generation of iPad 2. In order to help you decide which Apple tablet suits your requirements here is an infographic comparing the iPad  with iPad 2

    Read the article

  • How do you balance documentation requirements with Agile developments

    - by Jeremy
    In our development group there is currently discussions around agile and waterfal methodology. No-one has any practical experience with agile, but we are doing some reading. The agile manifesto lists 4 values: Individuals and interactions over processes and tools Working software over comprehensive documentation Customer collaboration over contract negotiation Responding to change over following a plan We are an internal development group developing applications for the consumption of other units in our enterprise. A team of 10 developers builds and releases multiple projects simultanously, typically with 1 - maybe 2 (rarely) developer on each project. It seems to be that from a supportability perspective the organization needs to put some real value on documentation - as without it, there are serious risks with resourcing changes. With agile favouring interactions, and software deliverables over processes and documentation, how do you balance that with the requirements of supportable systems and maintaining knowledge and understanding of how those systems work? With a waterfall approach which favours documentation (requirements before design, design specs before construction) it is easy to build a process that meets some of the organizational requirements - how do we do this with an agile approach?

    Read the article

  • HDMI audio output problems with Radeon card

    - by Matt Robinson
    No matter how much advice I follow, I still cannot get any audio to come out through my HDMI connection. I've tried downloading the latest proprietary FGLRX graphics drivers, and I've also gone into /etc/default/grub and altered GRUB_CMDLINE_LINUX_DEFAULT="quiet splash" into this GRUB_CMDLINE_LINUX_DEFAULT="quiet splash radeon.audio=1" But still I cannot get any sound nor can I get HDMI sound output to show up in Sound Settings or Pulse Audio Volume Control. I can get video to show up just fine on my monitor through HDMI so I'm sure this problem is fixable! I know this is an old problem with Ubuntu, but any new strategies into the problem would be much appreciated. With that being said, here are some of my specs: Ubuntu 12.04 LTS AMD Radeon HD 7520G

    Read the article

  • Video monetising and simple shop, platform?

    - by fieldman
    I have this client that wants a single-product website. The product is a training-video that they want to deliver virtually and optionally physically. I usually do all the front-end design and back-end development but the budget is close to $0 to start with. So I'm looking for a platform like shopify or something where a shop/cart can be set up quickly and simply with minimal up-front cost - but which can accomodate some kind of paywall (DRM too?) for the online video with an option to purchase for an aditional cost the physical DVD. Am I approaching the wrong way all together? Or do you know of any platform that will accomodate the specs?

    Read the article

  • kubuntu - what's the best smplayer configuration for best quality in hd movies (mkv)

    - by Frank
    I have ubuntu x64 13.04 with kde 4.11 and smplayer v0.8.6 and the last mplayer version from ppa. I have ATI video card HD6870 MSI with fglrx driver v13.4. My kde settings are: Composition mode: Opengl 3.1 graphic system qt: Raster scaling mode: precise Vsync: auto So what's the best configuration for quality over performace in smplayer according to my system specs? For example what do I have to set for the following options? enable postprocessing by default and postprocessing quality output driver Deinterlacing method software equalizer direct rendering double buffering draw video using slices threads for decoding (MPEG-1/2 and H.264 only loop filter use CoreAVC Thanks

    Read the article

  • Dual Screen with 12.04 ATI randr extension is not present

    - by Trevor Pearson
    Here's my problem I have two screens. One is a 22"led monitor and the other is 42 inch led tv. In windows 7 I run xbmc on the second monitor. I'm trying to mimic the same function in setting xbmc to display on the tv only. So I installed the latest x64 linux drivers from ATI. I configured (using the catalyst control panel) using single display (multi desktop) I then got a white screen with black x so I enable xinerama. with xinerama enable I got the displays to work correctly, however I received an error when I tried to enter display settings to change the launcher location. the error message was " "randr extension not present =" So I tried to install libxandr2 using terimal but here's what I get trev@Lrig:~$ sudo get-apt install libxander2 [sudo] password for trev: sudo: get-apt: command not found I'm at a loss now because I can't find a solution for the xandr error message. I'f my specs are important amd athx2 6400+ ghz 8 gigs ram radeon hd6950

    Read the article

  • Kernel panic: "Machine check: processor context corrupt" after install

    - by Red
    I am new to Linux and I am having issues installing. I have tried multiple versions and all seem to give me the same issue. The LiveCD worked and I used it to install Ubuntu alongside Windows 7. Here's a few screen shots of what I get. I am currently trying to install most the most recent version of Ubuntu. System specs: x58 motherboard 12GB RAM Core i7 processor 2 Nvidia GTX 580 video cards in SLI configuration Thanks in advance and please feel free to simplify your answers as I am pretty new at this.

    Read the article

  • Why do people think SOAP is deprecated?

    - by user98q37479
    While browsing SO today I found this question here and it starts with this: Sure, you're gonna tell me that SOAP is depracated and all, well i'm forced to use it Found lots of statement like this one on SO up till now, this one just triggered me to ask this question. REST has its uses, SOAP has its uses, in some places they intersect as functionality but they are not replaceable to one another. So I wonder, why do people think SOAP is "deprecated"? Is it ignorance? Complexity of SOAP and WS-* specs? REST hype? What? If you think SOAP is deprecated please tell me why. I'm curious!

    Read the article

  • Why do webpages take longer to loo in ubuntu 12.04 than Windows 7

    - by Emil Abraham
    For example, when I click on a Facebook picture, the picture remains pixelated for about 30 seconds, then starts to clear up. Or when I watch YouTube videos, I can't watch them on HD without running into buffer issues. Windows 7 is just much snappier. It might be an issue with the graphics card. Dualbooting Windows 7 64bit & Ubuntu 12.04 64 bit Specs: CPU: Intel® Core™ i7-2630QM CPU @ 2.00GHz × 8 RAM: 8GB DDR3 HDD: 50 GB to Ubuntu & Remaining 1.5 TB to Windows The interesting part: Graphics Card: On System Settings in Ubuntu: Intel® Sandybridge Mobile Graphics Card: What it should be: Radeon™ HD 7690M XT switchable graphics with 1024MB GDDR5 and up to 5093MB total graphics memory

    Read the article

  • Choppy video in vlc even with hardware acceleration on ATI Radeon HD

    - by l0b0
    After installing Catalyst 11.12 and studying instructions (1, 2) to enable hardware acceleration I ended up with the following command: sudo apt-get install xvba-va-driver libva-glx1 vainfo && \ vainfo && \ vlc --verbose 2 --ffmpeg-hw video.1080p.x264.mkv The output confirms that it's using the VA API: avcodec decoder: Using VA API version 0.32 for hardware decoding. But even after a reboot the video still looks choppy. There's no vlc output to indicate that the codec is having trouble, or that it's skipping frames. Any idea what could be the problem? Using vlc 1.1.12 with default settings. Also tried (in sequence): sudo apt-get install libva-glx1 sudo apt-get install --reinstall libgl1-mesa-glx:i386 libgl1-mesa-glx:amd64 libgl1-mesa-dri:i386 libgl1-mesa-dri:amd64 xserver-xorg-core Rebooting No luck yet. Specs from /proc/cpuinfo : Intel Core i7 CPU 950 @ 3.07GHz

    Read the article

  • How many developers before continuous integration becomes effective for us?

    - by Carnotaurus
    There is an overhead associated with continuous integration, e.g., set up, re-training, awareness activities, stoppage to fix "bugs" that turn out to be data issues, enforced separation of concerns programming styles, etc. At what point does continuous integration pay for itself? EDIT: These were my findings The set-up was CruiseControl.Net with Nant, reading from VSS or TFS. Here are a few reasons for failure, which have nothing to do with the setup: Cost of investigation: The time spent investigating whether a red light is due a genuine logical inconsistency in the code, data quality, or another source such as an infrastructure problem (e.g., a network issue, a timeout reading from source control, third party server is down, etc., etc.) Political costs over infrastructure: I considered performing an "infrastructure" check for each method in the test run. I had no solution to the timeout except to replace the build server. Red tape got in the way and there was no server replacement. Cost of fixing unit tests: A red light due to a data quality issue could be an indicator of a badly written unit test. So, data dependent unit tests were re-written to reduce the likelihood of a red light due to bad data. In many cases, necessary data was inserted into the test environment to be able to accurately run its unit tests. It makes sense to say that by making the data more robust then the test becomes more robust if it is dependent on this data. Of course, this worked well! Cost of coverage, i.e., writing unit tests for already existing code: There was the problem of unit test coverage. There were thousands of methods that had no unit tests. So, a sizeable amount of man days would be needed to create those. As this would be too difficult to provide a business case, it was decided that unit tests would be used for any new public method going forward. Those that did not have a unit test were termed 'potentially infra red'. An intestesting point here is that static methods were a moot point in how it would be possible to uniquely determine how a specific static method had failed. Cost of bespoke releases: Nant scripts only go so far. They are not that useful for, say, CMS dependent builds for EPiServer, CMS, or any UI oriented database deployment. These are the types of issues that occured on the build server for hourly test runs and overnight QA builds. I entertain that these to be unnecessary as a build master can perform these tasks manually at the time of release, esp., with a one man band and a small build. So, single step builds have not justified use of CI in my experience. What about the more complex, multistep builds? These can be a pain to build, especially without a Nant script. So, even having created one, these were no more successful. The costs of fixing the red light issues outweighed the benefits. Eventually, developers lost interest and questioned the validity of the red light. Having given it a fair try, I believe that CI is expensive and there is a lot of working around the edges instead of just getting the job done. It's more cost effective to employ experienced developers who do not make a mess of large projects than introduce and maintain an alarm system. This is the case even if those developers leave. It doesn't matter if a good developer leaves because processes that he follows would ensure that he writes requirement specs, design specs, sticks to the coding guidelines, and comments his code so that it is readable. All this is reviewed. If this is not happening then his team leader is not doing his job, which should be picked up by his manager and so on. For CI to work, it is not enough to just write unit tests, attempt to maintain full coverage, and ensure a working infrastructure for sizable systems. The bottom line: One might question whether fixing as many bugs before release is even desirable from a business prespective. CI involves a lot of work to capture a handful of bugs that the customer could identify in UAT or the company could get paid for fixing as part of a client service agreement when the warranty period expires anyway.

    Read the article

  • How do you cope with ugly code that you wrote?

    - by Ralph
    So your client asks you to write some code, so you do. He then changes the specs on you, as expected, and you diligently implement his new features like a good little lad. Except... the new features kind of conflict with the old features, so now your code is a mess. You really want to go back and fix it, but he keeps requesting new things and every time you finish cleaning something, it winds up a mess again. What do you do? Stop being an OCD maniac and just accept that your code is going to wind up a mess no matter what you do, and just keep tacking on features to this monstrosity? Save the cleaning for version 2?

    Read the article

  • XBMC freezes pc after viewing movies

    - by Eric
    Tough to replicate, but basically after watching any movie in my library on XBMC, it'll randomly start freezing my pc. All I can do is either force shutdown by holding down the power button or wait for it to shut itself down. It's not any particular movie, it's not any particular amount of time watching one. It only happens while using XBMC. PC is older, I know the hardware is outdated, but it's not overheating nor is a simple movie too much for it to play, at least I don't think it is. Specs: Intel Core 2 6400. nVidia Quardro FX 560 4GB RAM So any suggestions? I don't have any bug/error reports because it doesn't print out any. I'm trying to find a program like XBMC, as I like having my movies organized and not have to hunt through folders to find them, so any suggestions there?

    Read the article

  • How can I avoid the random restart of the xserver?

    - by Bernd
    I'm using a desktop pc with 64bit Ubuntu 12.04 (Kernel 3.2.0-24-generic). Hardware specs are - Intel Core i7 CPU 860 @ 2,80GHz x 8 - Nvidia GeForce GTS 250 - 750 GB Hard Disk ATA WDC WD7501AALS-00E3A0 (for my /home partition) - 128 GB Solid-State Disk ATA PLEXTOR PX-128M2S (for all other partitions) Since I reinstalled the PC with Ubuntu 12.04 the xserver restarts randomly. Most times when I watch a video in the browser (maybe a flash issue?) but sometimes the restart/crash appears when I'm working in an text-editor. How can I locate the problem? Which information is needed for a useful answer?

    Read the article

  • I can't get suspend, hibernate and shutdown to work in Ubuntu 12.04

    - by Fostext
    I recently built a computer with these specs: Asus Motherboard, Intel i3 3.3 GHz dual processor, 8 GB of RAM. I installed 12.04 on a brand new hard drive. I partitioned the hard drive between root, home and swap like I have often read how to do. I cannot get this machine to properly shutdown. I have to hold the power button down now. Although, for the first few days it properly shutdown. I also cannot get the system to hibernate or suspend properly. I have read tons on this and watched many YouTube tutorials on how to fix both, but the computer never wakes up after suspend or hibernate. It just stays on a black screen. Can anyone help? I love 12.04 so far, but these setbacks are making me worried about stability and power management issues. Also, I wonder if it's really bad for the hard drive to force the CPU to shutdown.

    Read the article

< Previous Page | 9 10 11 12 13 14 15 16 17 18 19 20  | Next Page >