Search Results

Search found 8744 results on 350 pages for 'yann core'.

Page 38/350 | < Previous Page | 34 35 36 37 38 39 40 41 42 43 44 45  | Next Page >

  • Forgot the username

    - by prithviraj
    Hello all I have fedora installed in my system. I know the password but i forgot the user name. I can access through terminal but i don't no how to login through gui. Please help me. Thanks in advance.

    Read the article

  • Processor Upgrade HP Elite M9510F

    - by DaveM
    I have an HP 9510F that uses the ASUS IPIBL-LB MB. It ships with an Intel Q8200 Quad Core processor but it does not support virtualization. Specs for the board from HP (ASUS does not list this OEM board) do not show support for the Intel Q8200 it ships with (obviously incorrect) but only these • Supports the following processors: o Intel Core 2 Quad (Yorkfield core) Q9xxx o Intel Core 2 Duo (Wolfsdale core) E8xxx o Intel Core 2 Quad (Kentsfield core) up to Q6600 o Core 2 Duo E6x00 (Conroe core) up to E6700 o Core 2 Duo E4x00 (Conroe core) up to E4400 Can this MB support the Q8400 or will it only support the indicated Q9xxx series? Naturally HP is little help here. Specs are located hereHP/ASUS MB specs

    Read the article

  • ffmpeg not using all cores

    - by user2783132
    I just got my new server with two intel e5-2695 but I was shocked to see that ffmpeg or ubuntu doesn't utilize all cores. top while ffmpeg was running top - 23:35:25 up 2:41, 2 users, load average: 5.35, 4.37, 3.12 Tasks: 333 total, 2 running, 331 sleeping, 0 stopped, 0 zombie %Cpu0 : 0.0 us, 1.0 sy, 35.6 ni, 63.4 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu1 : 0.0 us, 0.7 sy, 35.5 ni, 63.9 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu2 : 0.0 us, 0.7 sy, 33.4 ni, 65.9 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu3 : 0.0 us, 0.0 sy, 32.7 ni, 67.3 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu4 : 0.0 us, 0.3 sy, 32.3 ni, 67.3 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu5 : 0.0 us, 0.3 sy, 33.0 ni, 66.7 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu6 : 0.0 us, 0.0 sy, 32.6 ni, 67.4 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu7 : 0.0 us, 0.3 sy, 32.7 ni, 67.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu8 : 0.0 us, 0.7 sy, 32.6 ni, 66.8 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu9 : 0.0 us, 0.3 sy, 33.9 ni, 65.8 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu10 : 0.0 us, 0.0 sy, 35.0 ni, 65.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu11 : 0.0 us, 0.7 sy, 30.0 ni, 69.3 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu12 : 21.1 us, 0.0 sy, 0.0 ni, 78.9 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu13 : 0.7 us, 0.0 sy, 4.3 ni, 95.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu14 : 0.3 us, 0.0 sy, 5.0 ni, 94.6 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu15 : 24.9 us, 0.0 sy, 0.0 ni, 75.1 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu16 : 0.3 us, 0.0 sy, 3.7 ni, 96.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu17 : 0.7 us, 0.3 sy, 4.9 ni, 94.1 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu18 : 1.0 us, 0.0 sy, 4.6 ni, 94.4 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu19 : 0.7 us, 0.0 sy, 4.7 ni, 94.7 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu20 : 11.1 us, 0.0 sy, 0.0 ni, 88.9 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu21 : 1.3 us, 0.0 sy, 4.6 ni, 94.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu22 : 2.0 us, 0.3 sy, 4.3 ni, 93.4 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu23 : 96.7 us, 1.0 sy, 0.0 ni, 2.3 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu24 : 0.0 us, 0.0 sy, 0.7 ni, 99.3 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu25 : 0.0 us, 0.0 sy, 3.0 ni, 97.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu26 : 0.0 us, 0.0 sy, 1.3 ni, 98.7 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu27 : 0.0 us, 0.0 sy, 4.0 ni, 96.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu28 : 0.0 us, 0.0 sy, 1.7 ni, 98.3 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu29 : 0.0 us, 0.0 sy, 1.7 ni, 98.3 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu30 : 0.0 us, 0.0 sy, 1.7 ni, 98.3 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu31 : 0.0 us, 0.0 sy, 1.0 ni, 99.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu32 : 0.0 us, 0.0 sy, 0.7 ni, 99.3 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu33 : 0.0 us, 0.0 sy, 1.7 ni, 98.3 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu34 : 0.0 us, 0.0 sy, 2.0 ni, 98.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu35 : 0.0 us, 0.0 sy, 1.0 ni, 99.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st %Cpu36 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, 0.0 st ffmpeg was sent with -threads 0 *I also tried sending ffmoeg with -threads 500- no difference

    Read the article

  • Need Windows XP VGA driver for i3 Haswell

    - by AFH
    Background: I have recently upgraded my hardware because the previous Pentium system started failing to the point that it would not run long enough to boot. It was obviously a hardware fault, but I had no way of knowing whether it was in the motherboard, CPU or memory. Not all the components were now available, so I decided to replace all three. In order to get some benefit from the expenditure, I though I would put in faster components, and for future-proofing went for recently released ones: MSI Z87-G41-PCMate and Intel i3-4130 with 4400 HD graphics. The system performs excellently with Ubuntu 13.10, so I know there are no hardware problems, but I need to continue running XP because it runs several thousand pounds (UK) worth of software, which meets my needs more than adequately: in some cases there is no longer support for later Windows releases, and in most others an expensive and to me unnecessary upgrade is required. Problem: The motherboard specifications claim Windows XP support for the live driver update utility, which misled me into believing that XP drivers were available. Not true: Intel have apparently refused to provide XP drivers for Haswell chips. The update program runs on XP, but finds no suitable Intel drivers. The system is more or less running on the default fail-safe VGA driver, but DirectX will not load, which stops a number of my applications from running. I have been trawling the internet for a month now, but I have not found a graphics driver which will load successfully: all show "This device cannot start. (code 10)". I don't need HDMI support: my monitor is 1280x1024 and connected through the VGA port, so all I need is a driver which will handle this resolution well enough to support DirectX. Has anyone found a driver which will do this? Please don't reply with information found from internet searches, unless you have actually solved this problem: be assured that I have been all round the houses looking at solutions which others have reported as working, but none of them does for me. Incidentally, I did find an Intel HD sound driver which XP accepts (winxp_145111.exe from Intel), though without connecting to an HDMI port on a TV or sound system I have no idea if it works in practice. However, the graphics section of the same driver fails, like all the others I've tried.

    Read the article

  • How to get Windows Server 2008 VM to use multiple cores

    - by David Fraser
    I have a Windows Server 2008 machine running in VirtualBox. On initial installation, only one processor was made available, but now I want to run it as a multiprocessor machine. I have made all four cores available in the VirtualBox settings (as well as enabling VT-x/AMD-V and Nested Paging), but Task Manager still only shows one CPU. However, the four CPU cores are visible in Device Manager under Processors. In the event log on startup, I can see the following relevant events: EventLog.6009 Microsoft (R) Windows (R) 6.00.6002 Service Pack 2 Multiprocessor Free Kernel-Processor-Power.4 Processor 0 exposes the following: 1 idle state(s), 0 performance state(s), 0 throttle state(s) Kernel-Processor-Power.4 Processor 255 exposes the following: 0 idle state(s), 0 performance state(s), 0 throttle state(s) Kernel-Processor-Power.4 Processor 255 exposes the following: 0 idle state(s), 0 performance state(s), 0 throttle state(s) Kernel-Processor-Power.4 Processor 255 exposes the following: 0 idle state(s), 0 performance state(s), 0 throttle state(s) How can I make this system actually boot up as a multiprocessor machine?

    Read the article

  • AMD switchable graphics are not working OR I don't know how to make them work

    - by Deus Deceit
    I don't know if this is the right place for this question, but I'll give it a shot. I have a Dell Inspiron 17R 5721. It's supposed to be using switchable graphics. It has an Intel HD 4000 and a Radeon HD 8730M and I'm using windows 8. My problem is this, I Installed the drivers that dell gives me but I don't see the AMD graphics card ANYWHERE ( I do see it in device manager but not anywhere else to select and play a game using AMD). I installed the latest drivers from AMD and same thing, I can't run a game with AMD graphics card. I change the applications preferences in the Catalyst control center but even after that, games don't give me the option to select the AMD card, they list only the Intel HD 4000. Can someone tell me what I have to do to make this work? ------------------- After looking around and messing with stuff. I think.... I THINK that you don't really get the option to select a graphics card. Switchable graphics is all about switching automatically depending on application's needs. Cause when I uninstalled AMD's drivers or actually (screwed up lol) games were playing much worse. When I re-installed them, games went back to being good looking. So even if a game sees only the Intel HD 4000 graphics card windows or AMD's drivers will switch to the AMD Readeon graphics card automatically. I hope someone can verify this. Cause seriously I don't think you get to play scyrim with High graphics settings or even ultra with the Intel HD Graphics card. -------------------

    Read the article

  • intel HD graphics with integrated tv tuner

    - by Tamir
    Hi all! I have new Dell laptop with Intel HD graphics with integrated TV tuner. How can i use this TV tuner? should I install third party software for using it or just configure something? I tried to google it but couldn't find a thing :-( so, How can I use this TV tuner? Many thanks!

    Read the article

  • How to get Windows Server 2008 VM to use multiple cores

    - by David Fraser
    I have a Windows Server 2008 machine running in VirtualBox. On initial installation, only one processor was made available, but now I want to run it as a multiprocessor machine. I have made all four cores available in the VirtualBox settings (as well as enabling VT-x/AMD-V and Nested Paging), but Task Manager still only shows one CPU. However, the four CPU cores are visible in Device Manager under Processors. In the event log on startup, I can see the following relevant events: EventLog.6009 Microsoft (R) Windows (R) 6.00.6002 Service Pack 2 Multiprocessor Free Kernel-Processor-Power.4 Processor 0 exposes the following: 1 idle state(s), 0 performance state(s), 0 throttle state(s) Kernel-Processor-Power.4 Processor 255 exposes the following: 0 idle state(s), 0 performance state(s), 0 throttle state(s) Kernel-Processor-Power.4 Processor 255 exposes the following: 0 idle state(s), 0 performance state(s), 0 throttle state(s) Kernel-Processor-Power.4 Processor 255 exposes the following: 0 idle state(s), 0 performance state(s), 0 throttle state(s) How can I make this system actually boot up as a multiprocessor machine?

    Read the article

  • Computer freezes for 2+ seconds, mouse still moves

    - by xsaero00
    I have this problem on my workstation. The computer would effectively freeze for 2-5 seconds for no apparent reason, then continue as normal. While frozen the mouse would still be movable, but only on one of screens in my multi-screen setup. What can be the likely cause. System: CPU: i7-920 Memory: 12G of Patriot DDR3, 6 modules OS: SLED 11, Suse Linux Enterprise Desktop, using Gnome Main board: Asus P6T Video: two Nvidia 9500GT connected to three displays I am using memory at recommended settings of 8-8-8-1333. It has an XMP profile. Th CPU is a bit overclocked to 3.3 GHz, but my cooling more than allows for it. I ran the computer with all overclocks off and lower memory speed but the issue was still there. Any ideas? Where should I start looking?

    Read the article

  • Display with intel integrated graphics, bitcoin mine with Radeon 6950

    - by karategeek6
    I'm on Ubuntu Linux 11.04 64 bit. I have an intel i5 with integrated graphics and a Radeon 6950, with one monitor. I would like to run my graphics on the integrated card, and run bitcoin mining on the 6950. I have bitcoin mining working when I use the 6950 for both display and mining. Every time I try and and use the integrated graphics instead, OpenCL doesn't recognize my 6950. Using aticonfig --initial when using the integrated graphics for display breaks things. So I used the xorg.conf it created as a basis and tried to manually edit it. I really don't know what I'm doing, though. My last attempt is given below. The graphics ran off the integrated card, but the 6950 wasn't recognized. Any help would be greatly appreciated! xorg.conf: #Section "ServerLayout" # Identifier "Intel Layout" # Screen "Default Screen" # Identifier "aticonfig Layout" # Screen "aticonfig-Screen[0]-0" # Screen 0 "aticonfig-Screen[0]-0" 0 0 #EndSection Section "Module" Load "glx" EndSection # Intel Section "Device" Identifier "Intel Integrated Graphics" Driver "intel" BusID "PCI:0:2:0" EndSection Section "Monitor" Identifier "Default Monitor" Option "VendorName" "Monitor Vendor" Option "ModelName" "Monitor Name" Option "DPMS" "true" EndSection Section "Screen" Identifier "Default Screen" Device "Intel Integrated Graphics" Monitor "Default Monitor" DefaultDepth 24 EndSection # ATI Section "Device" Identifier "aticonfig-Device[0]-0" Driver "fglrx" BusID "PCI:1:0:0" EndSection Section "Monitor" Identifier "aticonfig-Monitor[0]-0" Option "VendorName" "ATI Proprietary Driver" Option "ModelName" "Generic Autodetecting Monitor" Option "DPMS" "true" EndSection Section "Screen" Identifier "aticonfig-Screen[0]-0" Device "aticonfig-Device[0]-0" Monitor "aticonfig-Monitor[0]-0" DefaultDepth 24 SubSection "Display" Viewport 0 0 Depth 24 EndSubSection EndSection

    Read the article

  • Nvidia 9 series or Intel HD 2000? [closed]

    - by EApubs
    I just tested an Nvidia 9300 GS card with a Intel Corei3 HD 2000 graphics system. Here is the windows experience index scores I got : Nvidia 9300 GS : Base Score 3.9 Processor : 7.1 Memory : 7.5 Graphics : 3.9 Gaming Graphics : 5.1 Hard Disk : 5.9 Intel HD 2000 : Base Score : 5.2 Processor : 7.1 Memory : 5.9 Graphics : 5.2 Gaming Graphics : 5.8 Hard Disk : 5.9 My questions are : When using Intel HD graphics, it reduces the score of my Ram! How is that possible? It checks the speed of the ram. Not the size (i think). Intel graphics take some of the ram space but how can that effect the speed? From both of them, what will be the good choice?

    Read the article

  • PC wont boot with more than 1 stick of RAM.

    - by Aidan
    Hi Guys, I've got the following computer and I've just put in a new CPU QX9650 and I've run into this problem since making this hardware change. Whenever I put more than 1 of my 4 sticks of ram into my machine it wont load an OS. It'll go through the BIOS but BSOD on windows load. It also wont let me install an OS from disk or boot into Linux. I've ran memtest with all 4 sticks in and I get 10k+ errors on test5. Each stick of ram on it's own is fine and functions properly. I only have problems when all 4 sticks are in the machine at the same time. System specs.. CPU: QX9650 Mobo: Asus P5B 2104 BIOS RAM: 2xPC25400 DDR2 , 2xPC2 6400 both OCZ. Is the problem on my end or is the CPU faulty?

    Read the article

  • PC wont boot with more than 1 stick of RAM.

    - by Aidan
    Hi Guys, I've got the following computer and I've just put in a new CPU QX9650 and I've run into this problem since making this hardware change. Whenever I put more than 1 of my 4 sticks of ram into my machine it wont load an OS. It'll go through the BIOS but BSOD on windows load. It also wont let me install an OS from disk or boot into Linux. I've ran memtest with all 4 sticks in and I get 10k+ errors on test5. Each stick of ram on it's own is fine and functions properly. I only have problems when all 4 sticks are in the machine at the same time. System specs.. CPU: QX9650 Mobo: Asus P5B 2104 BIOS RAM: 2xPC25400 DDR2 , 2xPC2 6400 both OCZ. Is the problem on my end or is the CPU faulty?

    Read the article

  • kill SIGABRT does not generate core file from daemon started from crontab.

    - by Guma
    I am running CentOS 5.5 and working on server application that sometimes I need to force core dump so I can see what is going on. If I start my server from shell and send kill SIGABRT, a core file is created. If I start same program from crontab and then I send the same signal to it the server is "killed" but no core file is generated. Does any one know why is that and what need to be added to my code or changed in system settings to allow core file generation? Just a side note I have ulimit set to unlimited in /etc/profile I have set kernel.core_uses_pid = 1 kernel.core_pattern=/var/cores/%h-%e-%p.core in /etc/sysctl.conf Also my server app was added to crontab under same login id as I am running it from shell. Any help greatly appreciated

    Read the article

  • Is there any way to limit the turbo boost speed / intensity on i7 lap?

    - by Anonymous
    I've just got a used i7 laptop, one of these overheating pavilions from HP with quad cores. And I really want to find a compromise between the temp and performance. If I use linpack, or some other heavy benchmark, the temp easily gets to 95+, and having a TJ of 100 Degrees, for a 2630QM model, it really gets me throttling, that no cooling pad or even an industrial fan could solve. I figured later that it is due to turbo boost, and if I set my power settings to use 99% of the CPU instead of 100%, and it seems to disable the turbo boost, so the temp gets better. But then again it loses quite a bit of performance. The regular clock is 2GHz, and in turbo boost it gets to 2.6Ghz, but I just wonder if I could limit it to around 2.3Ghz, that would be a real nice thing. Also there is another question I've hard time getting answer to. It seems to me that clocks are very quickly boosting up to max even when not needed, eg, it's ok if the CPU has 0% load, the clocks get to their 800MHz, but even if it gets to about 5% it quickly jumps to a max and even popping up turbo, which seems very strange to me. So I wonder if there is any way to adjust the sensitivity of the Speed Step feature. I believe it would be more logical to demand increased clock if it hits let's say 50% load. I do understand that most of these features are probably hardwired somewhere in the CPU itself or the MB, which has no tuning options just like on many laptops. But I would appreciate if you could recommend some thing, or some software. Thanks

    Read the article

  • SQL server availability issue: large query stops other connections from connecting

    - by Carlos
    I've got a high-spec (multicore, RAID) server running MS SQL 2008, with several databases on it. I have a low throughput process that periodically needs a small amount of information from one of the DBs, and the code seems to work fine. However, sometimes when one of my colleagues does a huge query against one of the other DBs, I see full CPU usage on the machine, and connections from my app time out. Why does this happen? I would have thought the many cores and harddisks would somehow (together with cleverly written DB server) be able to keep at least some of the resources free for other apps? I'm pretty sure he doesn't use multiple connections for his query. What can I do to prevent this?

    Read the article

  • i7-980X at 70% speed

    - by Buxley
    Hi we bought a nice computer to use to solve optimization problems. Intel i7-980X@3,33 GHz with 12 GB of Team Group 1600 MHz DDR3 RAM. When we use Gurobi the computer uses all 12 cores at maximum in the beginning of the solve. However after a while (about 8 hrs) it all cores jump between 65 and 85% When I solve the same models on an i7 930 all cores are at a near 100% level even after longer solution times. We first thought that the Hard disk was the bottleneck since Gurobi writes out nodefiles after the memorylimit is used. However since the new computer have 12 GB of RAM we put the memorylimit to 7 GB so the solver only used the RAM and still with the same performance in the processor. Any ideas about the bottleneck? As I said earlier it works at 100% for the first hours or so . Thanks very much for any answers! Our plan was to overclock it but we can't even get it to work at normal speed yet!

    Read the article

  • i7 x980 at 70% speed

    - by Buxley
    Hi we bought a nice computer to use to solve optimization problems. intel i7 X980@3,33 Mhz with 12 Gb of Team Group 1600 MHz ddr3 Ram. When we use Gurobi The Computer uses all 12 cores at maximum in the beginning of the solve. However after a while (about 8 hrs) it all cores jump between 65 and 85% When I solve the same models on an I7 930 all cores are at a near 100% level even after longer solution times. we first thought that the Harddisk was the bottleneck since Gurobi writes out nodefiles after the memorylimit is used. However since the new computer have 12 GB of Ram we put the memorylimit to 7 GB so the solver only used the RAM and still with the same performance in the processor. Any Ideas about the bottleneck? As I said earlier it works at 100% for the first hours or so . Thanks very much for any answers! Our plan was to overclock it but we can't even get it to work at normal speed yet!

    Read the article

  • What is good RAM for my CPU?

    - by Jason94
    I'm going to upgrade my RAM, but I have no clue what to look for. There is cheap and there is expensive RAM. What should I look for? What are good attributes for intel i5? The i5 CPU and motherboeard supports 1600 MHz (PC3-12800), but there is still to much to choose from. There are CL10 (10-10-10-27) @ 1.5V CL11 @ 1.35V CL9 (9-9-9-24) @ 1.5V CL9 (9-9-9-24) @ 1.65V Higher CL is better? CL is classification? And I guess a high voltage is also good since thats what overclockers do to the ram (increase the voltage)?

    Read the article

  • How do I fix these LibreOffice unmet dependencies?

    - by Lucky
    The following packages have unmet dependencies: libreoffice: Depends: libreoffice-core (= 1:3.4.4-0ubuntu1) but 1:3.4.3-3ubuntu2 is to be installed Depends: libreoffice-base but it is not going to be installed Depends: libreoffice-report-builder-bin but it is not going to be installed Depends: ttf-dejavu but it is not going to be installed Depends: ttf-sil-gentium-basic but it is not going to be installed Depends: libreoffice-filter-mobiledev but it is not going to be installed Depends: libreoffice-java-common (>= 1:3.4.4~) but it is not going to be installed libreoffice-base-core : Depends: libreoffice-core (= 1:3.4.4-0ubuntu1) but 1:3.4.3-3ubuntu2 is to be installed libreoffice-calc : Depends: libreoffice-core (= 1:3.4.4-0ubuntu1) but 1:3.4.3-3ubuntu2 is to be installed libreoffice-draw : Depends: libreoffice-core (= 1:3.4.4-0ubuntu1) but 1:3.4.3-3ubuntu2 is to be installed libreoffice-gnome : Depends: libreoffice-core (= 1:3.4.4-0ubuntu1) but 1:3.4.3-3ubuntu2 is to be installed libreoffice-gtk : Depends: libreoffice-core (= 1:3.4.4-0ubuntu1) but 1:3.4.3-3ubuntu2 is to be installed libreoffice-impress : Depends: libreoffice-core (= 1:3.4.4-0ubuntu1) but 1:3.4.3-3ubuntu2 is to be installed libreoffice-math : Depends: libreoffice-core (= 1:3.4.4-0ubuntu1) but 1:3.4.3-3ubuntu2 is to be installed libreoffice-writer : Depends: libreoffice-core (= 1:3.4.4-0ubuntu1) but 1:3.4.3-3ubuntu2 is to be installed python-uno : Depends: libreoffice-core (= 1:3.4.4-0ubuntu1) but 1:3.4.3-3ubuntu2 is to be installed E: Unmet dependencies. Try 'apt-get -f install' with no packages (or specify a solution).

    Read the article

  • C++ SDL State Machine Segfault

    - by user1602079
    The code compiles and builds fine, but it immediately segfaults. I've looked at this for a while and have no idea why. Any help is appreciated. Thank you! Here's the code: main.cpp #include "SDL/SDL.h" #include "Globals.h" #include "Core.h" #include "GameStates.h" #include "Introduction.h" int main(int argc, char** args) { if(core.Initilize() == false) { SDL_Quit(); } while(core.desiredstate != core.Quit) { currentstate->EventHandling(); currentstate->Logic(); core.ChangeState(); currentstate->Render(); currentstate->Update(); } SDL_Quit(); } Core.h #ifndef CORE_H #define CORE_H #include "SDL/SDL.h" #include <string> class Core { public: SDL_Surface* Load(std::string filename); void ApplySurface(int X, int Y, SDL_Surface* source, SDL_Surface* destination); void SetState(int newstate); void ChangeState(); enum state { Intro, STATES_NULL, Quit }; int desiredstate, stateID; bool Initilize(); }; #endif Core.cpp #include "Core.h" #include "SDL/SDL.h" #include "Globals.h" #include "Introduction.h" #include <string> /* Initilizes SDL subsystems */ bool Core::Initilize() { //Inits subsystems, reutrns false upon error if(SDL_Init(SDL_INIT_EVERYTHING) == -1) { return false; } SDL_WM_SetCaption("Game", NULL); return true; } /* Loads surfaces and optimizes them */ SDL_Surface* Core::Load(std::string filename) { //The surface to be optimized SDL_Surface* original = SDL_LoadBMP(filename.c_str()); //The optimized surface SDL_Surface* optimized = NULL; //Optimizes the image if it loaded properly if(original != NULL) { optimized = SDL_DisplayFormat(original); SDL_FreeSurface(original); } else { //returns NULL upon error return NULL; } return optimized; } /* Blits surfaces */ void Core::ApplySurface(int X, int Y, SDL_Surface* source, SDL_Surface* destination) { //Stores the coordinates of the surface SDL_Rect offsets; offsets.x = X; offsets.y = Y; //Bits the surface if both surfaces are present if(source != NULL && destination != NULL) { SDL_BlitSurface(source, NULL, destination, &offsets); } } /* Sets desiredstate to newstate */ void Core::SetState(int newstate) { if(desiredstate != Quit) { desiredstate = newstate; } } /* Changes the game state */ void Core::ChangeState() { if(desiredstate != STATES_NULL && desiredstate != Quit) { delete currentstate; switch(desiredstate) { case Intro: currentstate = new Introduction(); break; } stateID = desiredstate; desiredstate = core.STATES_NULL; } } Globals.h #ifndef GLOBALS_H #define GLOBALS_H #include "SDL/SDL.h" #include "Core.h" #include "GameStates.h" extern SDL_Surface* screen; extern Core core; extern GameStates* currentstate; #endif Globals.cpp #include "Globals.h" #include "SDL/SDL.h" #include "GameStates.h" SDL_Surface* screen = SDL_SetVideoMode(640, 480, 32, SDL_SWSURFACE); Core core; GameStates* currentstate = NULL; GameStates.h #ifndef GAMESTATES_H #define GAMESTATES_H class GameStates { public: virtual void EventHandling() = 0; virtual void Logic() = 0; virtual void Render() = 0; virtual void Update() = 0; }; #endif Introduction.h #ifndef INTRODUCTION_H #define INTRODUCTION_H #include "GameStates.h" #include "Globals.h" class Introduction : public GameStates { public: Introduction(); private: void EventHandling(); void Logic(); void Render(); void Update(); ~Introduction(); SDL_Surface* test; }; #endif Introduction.cpp #include "SDL/SDL.h" #include "Core.h" #include "Globals.h" #include "Introduction.h" /* Loads all the assets */ Introduction::Introduction() { test = core.Load("test.bmp"); } void Introduction::EventHandling() { SDL_Event event; while(SDL_PollEvent(&event)) { switch(event.type) { case SDL_QUIT: core.SetState(core.Quit); break; } } } void Introduction::Logic() { //to be coded } void Introduction::Render() { core.ApplySurface(30, 30, test, screen); } void Introduction::Update() { SDL_Flip(screen); } Introduction::~Introduction() { SDL_FreeSurface(test); } Sorry if the formatting is a bit off... Having to put four spaces for it to be put into a code block offset it a bit. I ran it through gdb and this is what I got: Program received signal SIGSEGV, Segmentation fault. 0x0000000000400e46 in main () Which isn't incredibly useful... Any help is appreciated. Thank you!

    Read the article

  • Can anyone explain to me what problem Core Data solves?

    - by Curtis Sumpter
    Core Data seems to add a needless layer of complexity. If you want to save data created natively by the user in an app why not just use an object and then write the data all to SQLite or back to a server using a RESTful script if necessary. Android doesn't have Core Data (though if it has something similar I haven't seen it.). What the heck is the point of buggy CD except useless needless overhead for people who can't write SQL or CGI scripts?

    Read the article

< Previous Page | 34 35 36 37 38 39 40 41 42 43 44 45  | Next Page >