Search Results

Search found 28301 results on 1133 pages for 'external process'.

Page 235/1133 | < Previous Page | 231 232 233 234 235 236 237 238 239 240 241 242  | Next Page >

  • Distributing a very simple application

    - by vanna
    I have a very simple working console application written in C++ linked with a light static library. It is just for testing purposes. Now that the coding part is done, I would like to know the process of actually distributing the program. I wrote a very basic CMakeLists.txt that create makefiles or VS projects to build the sources. I also have a program that calls the static library in order to make some google tests. To me, the distribution of this application goes like this : to developpers : the src directory with the CMakeLists.txt file (multi-platform distribution) with a README.txt and an INSTALL.txt to users : the executable and a README.txt on my git repo : everything mentionned above plus the sources for testing and the gtest external lib A this point : considering the complexity of my application, am I doing it right ? Is there any reference that would formalize this distribution process so I can get better and go further ? Say I would like to add dynamic libraries that can be updated, external libraries like boost : how should I package this to distribute it in a professionnal way ?

    Read the article

  • PHP processes run one at a time, always taking 100% of one core

    - by Derek Kurth
    We have seven websites written in PHP running on a Windows 2008 server with IIS 7.5. They are all very slow right now. When I look in Task Manager, I see around 10 php-cgi.exe processes, and they are all taking 0% of the CPU, except one, which is taking 25%. It's a quad-core server, so it's taking 100% of one core. If I watch for a few seconds, the process taking 25% will go to 0%, and a different php-cgi.exe process will jump to 25%. So all the php-cgi.exe processes are just lined up, waiting on a single core, and each process uses 100% of the processor when it can. Each of the 7 sites is in its own application pool in IIS, and we're using FastCGI. The PHP version is 5.3. Any ideas? Thanks!

    Read the article

  • Managing PHP processes on Windows 7 (with WAMP)

    - by Andrea
    Is there a way to manage (especially list and kill) long-running PHP processes on a Windows 7 system set up with WAMP? Every once in a while, I'll accidentally throw an infinite loop into a PHP process and want to kill it. Right now, all I can think to do is to restart all my WAMP services but sometimes the PHP processes manage to survive right through the restart, i.e., I still see them outputting to logs even after WAMP's restarted. And if the process isn't logging, then I have no way at all to know when/if it's been killed. Not to mention, this will wipe out everything I'm doing with WAMP, not just a single process. I don't seem to see anything relevant in the Windows Task Manager, but maybe I'm missing something.

    Read the article

  • Rsyslog problem after ubuntu upgrade 10.4 to 12.4

    - by Oxymoron
    I was using Ubuntu 10.4 until last week for storing the log informations of a external device with rsyslog. After upgrading to ubuntu 12.4 the logging of TCP doesn't works anymore. (There are just no pakets visible - not even with tcpdump - aold ubuntu machine still sees the pakets.) UDP works with the identical configuration on the ubuntu machine and a "use UDP" on the external device. Are there any changes in rsyslog, that could explain this? My rsyslog.conf file looks like this (with more comments): $ModLoad imuxsock # provides support for local system logging $ModLoad imklog # provides kernel logging support (previously done by rklogd) #$ModLoad immark # provides --MARK-- message capability $KLogPath /proc/kmsg # provides UDP syslog reception $ModLoad imudp $UDPServerRun 514 # provides TCP syslog reception $ModLoad imtcp $InputTCPServerRun 514 ########################### #### GLOBAL DIRECTIVES #### ########################### $ActionFileDefaultTemplate RSYSLOG_TraditionalFileFormat # Set the default permissions for all log files. # $FileOwner syslog $FileGroup adm $FileCreateMode 0640 $DirCreateMode 0755 $Umask 0022 $PrivDropToUser syslog $PrivDropToGroup syslog if $fromhost-ip startswith '192.168.0.10' then /var/log/caliDevice.log & ~ # local/regular rules, like '.' /var/log/syslog.log $IncludeConfig /etc/rsyslog.d/*.conf

    Read the article

  • SEO: Is promoting your backlinks a good strategy for improving search results for my site's name?

    - by user4394
    I run a website that's been around for about three years in the sports space. I am successfully ranking well for targeted keywords, but searching for the name of my site itself returns very poor results - it shows my site, its FB/Twitter, and then 15 pages of unrelated spam that happen to contain two words that, when combined, form my website's name. After that, my backlinks begin to show up spordically. As far as I can tell, I simply don't have enough backlinks and the backlinks I do have are ranked worse than the spam. (Site Explorer lists 200 external links to any page on our domain and 20 external links directly to the front page). To counter this, my strategy is to promote my backlinks so they get a better page rank than the spam. Does that make sense? Am I going in the right direction or should I just focus on getting more backlinks pointing directly to my site? Thanks in advance and I'd be happy to answer any questions I can (without giving away my site of course).

    Read the article

  • Client/Server game even in solo: any big problem?

    - by Klaim
    I'm making a game which have strong basic design based on multiplayer but also should provide a really interesting and self-sufficient solo game. A bit like a real-time strategy game. The events and actions taken shouldn't be as massive and immediate as in a FPS, so you can also think the networking like for an RTS. It's a PC game, targetting Windows, MacOSX and Linux (Ubuntu & Fedora). It's programmed in C++, using a variety of open source libraries, so I have great (potential) control over the performances. So far I always considered that just making the game work with two applications, client & server, even in solo mode was ok. However, as I'm in the process of starting the network code I'm having doubts about if it's a good idea. I'm not a specialist so I might be missing something in my analysis. I see these pros and cons: Pros: The game works only one way so if I fix a bug it should apply on all game modes, whatever the distance with the server is; Basic networking issues would be detected early, including behaviour with the protection softwares (firewall) installed (i am not specialist so this might be wrong); Cons: I suppose that even if it should be really fast enough, networking client and server on the same computer would still be slower than no networking and message passing in (one) process memory. Maybe debugging would be more difficult? I don't have experience in this case but so far I assume that debugging with Visual Studio allows me to debug multiple process so it shouldn't be really different. Also, remote debugging. My question is: is there a big disadvantage that I missed? Or maybe there are advantages that I missed and that should encourage me to just continue with only client-server game sessions?

    Read the article

  • i18n and L10n (1)

    - by Aaron Li
    Internationalization (i18n) is a way of designing and developing a software product to function in multiple locales. This process involves identifying the locales that must be supported, designing features which support those locales, and writing code that functions equally well in any of the supported locales. Localization (L10n) is a process of modifying or adapting a software product to fit the requirements of a particular locale. This process includes (but may not be limited to) translating the user interface, documentation and packaging, changing dialog box geometries, customizing features (if necessary), and testing the translated product to ensure that it still works (at least as well as the original). i18n is a pre-requisite for L10n. Resource is 1. any part of a program which can appear to the user or be changed or configured by the user. 2. any piece of the program's data, as opposed to its code. Core product is the language independent portion of a software product (as distinct from any particular localized version of that product - including the English language version). Sometimes, however, this term is used to refer to the English product as opposed to other localizations.   Useful links http://www.mozilla.org/docs/refList/i18n/ http://www.w3.org/International/ http://hub.opensolaris.org/bin/view/Community+Group+int_localization/

    Read the article

  • Oracle Open World 2012 Call For Papers

    - by Lionel Dubreuil
    At Oracle OpenWorld, more than a thousand people demonstrate their mastery and expertise by leading sessions on a vast array of Oracle technologies and products.Now’s the time to submit your SOA Governance success story presentation abstract for review by the selection panel.Don’t delay—submit your abstract now as the Call for Papers is open through next Monday,  April 9th.The competition is strong: roughly 18% of entries are accepted each year from more than 3,000 submissions. Review panels are made up of experts both internal and external to Oracle. Successful submissions often (but not exclusively) focus on customer successes, how-tos, or technical topics.What’s in it for you? Recognition, for one thing. Accepted sessions are publicized in the content catalog, which goes live in mid-June, and sessions given by external speakers often prove the most popular.Plus, accepted speakers get a complimentary pass to Oracle OpenWorld with access to all sessions and networking events-that could save you up to $2,595!For more information, please look here.

    Read the article

  • Is there a way to force lubuntu screen resolution?

    - by za1
    I have a Dell Inspiron 4000, 900 Mhz Pentium III, 192 MB RAM,10.1 GB disk running Lubuntu 12.04. When I start the computer, and I check the display resolution, Monitor Settings claims that the max resolution is 1024x768. I then connected another monitor I had lying around, rebooted, and now the laptop monitor looks normal,and monitor settings claims that 1280x1024 is the max resolution (it is) and the other monitor doesn't turn on. (black screen) The external monitor is not broken. Is there a way to, without the external monitor, force the laptop monitor resolution to 1280x1024? (The following is another question) - How can you run commands automatically at startup? Specifically "xinput 18 118 0" (without quotes) The first question is important, I can live with typing the second one at every boot. Thanks

    Read the article

  • Booting off a ZFS root in 14.04

    - by RJVB
    I've been running a Debian derivative (LMDE) on a ZFS root for half a year now. It was created by cloning a regular ext4-based install with all the necessary packages onto a ZFS pool, chrooting into that pool and recreating a grub menu and bootloader. The system uses an ext-3 dedicated /boot partition. I would like to do the same with Ubuntu 14.04, but have encountered several obstacles. There is no Trusty zfs-grub package The default grub package doesn't have ZFS support built in. I found a small bug in the build system responsible for that (report with patch created) and built my own grub packages. The built-in ZFS support is dysfunctional, it does not add the proper arguments to the kernel command line I thus installed the ZoL grub package I also use on my LMDE system, which does give me a correct grub.cfg However, even with that correct grub.cfg, the boot process apparently doesn't retrieve the bootfs parameter from the ZFS pool; instead the variable that's supposed to receive the value remains empty. As a result, initrd tries to load the default pool ("rpool"), which fails of course. I can however import the pool by hand, and complete the process by hand. If memory serves me well, I also had to disable apparmor, to avoid the boot process from blocking after importing the pool. Am I overlooking something? Just for comparison, I installed the Ubuntu 3.13 kernel on my LMDE system, and that works just fine (i.e. the identical kernel and grub binaries allow successful booting without glitches on LMDE but not on Ubuntu).

    Read the article

  • How to cleanup tmp folder safely on Linux

    - by Syncopated
    I use RAM for my tmpfs /tmp, 2GB, to be exact. Normally, this is enough but sometimes, processes create files in there and fail to cleanup after themselves. This can happen if they crash. I need to delete these orphaned tmp files or else future process will run out of space on /tmp. How can I safely garbage collect /tmp? Some people do it by checking last modification timestamp, but this approach is unsafe because there can be long-running processes that still need those files. A safer approach is to combine the last modification timestamp condition with the condition that no process has a file handle for the file. Is there a program/script/etc that embodies this approach or some other approach that is also safe? Incidentally, does Linux/Unix allow a mode of file opening with creation wherein the created file is deleted when the creating process terminates, even if it's from a crash?

    Read the article

  • How do I log file system read/writes by filename in Linux?

    - by Casey
    I'm looking for a simple method that will log file system operations. It should display the name of the file being accessed or modified. I'm familiar with powertop, and it appears this works to an extent, in so much that it show the user files that were written to. Is there any other utilities that support this feature. Some of my findings: powertop: best for write access logging, but more focused on CPU activity iotop: shows real time disk access by process, but not file name lsof: shows the open files per process, but not real time file access iostat: shows the real time I/O performance of disk/arrays but does not indicate file or process

    Read the article

  • Inserting x200s into (ultrabase) docking station mirror screen is always activated leading to non optimal resolution

    - by kiu
    Builtin LCD should be 1440x900 External LCD should be 1920x1080 If X200s is inserted into docking station the option mirror screen is always activated leading to a resolution of 1152x864 which looks terrible on the builtin and external lcd. My manual configuration for docking mode (seperate screens with maximum resolution) should be respected, but "Make Default" button has no consequences. Found a quick fix, but this cant be the offical ubuntu way... /etc/udev/rules.d/99-vga.rules: SUBSYSTEM=="drm", ACTION=="change", RUN+="/usr/local/sbin/vga_changed.sh" /usr/local/sbin/vga_changed.sh: #!/bin/bash dmode="$(cat /sys/class/drm/card0-VGA-1/status)" export DISPLAY=:0.0 if [ "${dmode}" = disconnected ]; then /usr/bin/sudo -u kiu /usr/bin/xrandr --output LVDS1 --mode 1440x900 --pos 0x0 --output VGA1 --off elif [ "${dmode}" = connected ]; then /usr/bin/sudo -u kiu /usr/bin/xrandr --output LVDS1 --mode 1440x900 --pos 0x0 --output VGA1 --auto --mode 1920x1080 --right-of LVDS1 fi

    Read the article

  • Nvidia 8600 GT drivers Will not run setup

    - by Zeno
    I am attempting to upgrade my video card drivers with my nvidia 8600 gt. I downloaded the drivers from NVIDIA and when I run the setup, nothing happens. It does appear in the Processes, but never does anything. I've tried restarting, tried killing the process and re-running, I am an admin account... nothing works. What is wrong? [EDIT] The installer does run, it extracts the setup files to C:\NVIDIA as normal then when it tries to run that setup to do the actual install, nothing happens (the process does show up in Task Manager and just sits there). Attempting to run the installer manually via C:\NVIDIA\DisplayDriver\296.10\WinXP\English\Setup.exe has the process open, but nothing happens. This machine is WinXP 32bit.

    Read the article

  • Jumping Login Box after Lighdm Multiple Monitor workaround

    - by Tom Gamon
    So I used this workaround to sort my resolution at the login screen when using multiple monitors with Lightdm. #!/bin/bash XCOM0=`xrandr -q | grep 'VGA1 connected'` XCOM1=`xrandr --output LVDS1 --primary --auto --output VGA1 --auto --right-of LVDS1` XCOM2=`xrandr --output LVDS1 --primary --auto` # if the external monitor is connected, then we tell XRANDR to set up an extended desktop if [ -n "$XCOM0" ] || [ ! "$XCOM0" = "" ]; then echo $XCOM1 # if the external monitor is disconnected, then we tell XRANDR to output only to the laptop screen else echo $XCOM2 fi exit 0; Found Here: How to force Multiple Monitors correct resolutions for LightDM? It works great. However, now when I am on my login screen, the login box seems to jump to between the two displays. Any advice as to how I could make it stay on one display? Thanks

    Read the article

  • How to share/access to partition from ubuntu vmware

    - by chr
    I am beginner at Ubuntu. Here is my problem. I have Ubuntu installed on my external HDD and i am running XP through vmware on Ubuntu, because my internal disk is dead atm. External HDD have ext4 (37gb) and 2 NTFS partition (36gb and 220gb). My question is, how i can access that 220gb (or 36gb) NTFS partition from vmware XP? I was already try search for similar posts but no luck to solve my problems. Thank you in advance Regards

    Read the article

  • Reliance on Outlook (been a looong time, I know)

    - by AndyScott
    Do you feel that your development group too reliant on Outlook? Have you reached a point that you have to search your email for pertinent information when asked? What are you using? I realized things had gotten out of hand a couple weeks ago over a weekend. I was at my in-laws house (in the country, no PC/laptop, no internet connection; and I get an email on my phone that I needed to reply to, but I couldn't send without deleting items from my inbox/sent items/etc. Now mind you, I have rules set up to move stuff into folders, and files more than a month old are automatically moved to the PST; but generally don't manually move items to a PST until I have had a chance to 'work' the item. Please don't bother mocking my process, it's just the way I work. That being said, it was a frustrating process of 'I need all this information, what can I afford to lose'. I work on an International project (think lots of customers), and conversations in 9 or 10 different directions about 10-20 different things are not abnormal for a given day. I have found myself looking data up in Outlook because that's where it is. I think that I have reached the point now, where I don't feel that Outlook is up to the task of organizing the data that it contains.   When you have that many emails (200 or so a day), information seems to get lost at times, and I find that Outlook's search capabilities are lacking. Additionally, I find that any sort of organizational 'system' of sorting emails that can cover multiple topics is a lost cause. But at the same time, the old process of taking the information that I got from emails and moving it into another 'notes' type of program has proved to be too time consuming. Anyone out there have some better type of system? (Comments about the capacity of my brain, and it's ability to recall information not needed.)

    Read the article

  • Installing 12.04 through Update Manager on a XP/ubuntu dual-boot

    - by Madeline Mcormick
    I currently have a dual-boot system running XP Pro SP3 with Ubuntu 10.04 LTS. I decided to upgrade to 12.04 using the Update Manager from the network and NOT using ISO CD version. Now that I am in the middle of 12.04 installation, I have this immense fear that this upgrade from update manager on the network server may affect my Win XP OS and may render it un-bootable. I tried backing up files while its upgrading to Ubuntu but it does not recognize any external media like external HDD. What should I do?

    Read the article

  • Spambot Infection Detection

    - by crankshaft
    My server has been blocked by CBL for participating in curtwail spambot. Initially we suspected that it was coming from a PC and not from the server, but the router is blocking all packets on 25 except those coming from the server. I have just executed the tcpdump command and every 5 minutes I see a flurry of activity on port 25 that is very suspicious and I am sure that there is some process running on the server: 13:02:30.027436 IP exprod5og110.obsmtp.com.53803 > ubuntu.local.smtp: Flags [S], seq 171708781, win 5744, options [mss 1436,sackOK,TS val 3046699707 ecr 0,nop,wscale 2], length 0 I have stopped postfix, and yet there is still traffic on port 25 above. But how can I find what process is actually communicating on port 25 as it only rund for a few seconds and so for example lsof -i :25 will never catch it. I have been working on this now for 2 days, it is a live server and I cannot simply shut it down, any suggestion on how I can detect the source of this email bot process ?

    Read the article

  • VMware Player 5.0 or VMware Workstation 9.0 after upgrade to Ubuntu 12.10

    The upgrade process Upgrading Ubuntu 12.04 to latest version 12.10 - aka Quantal Quetzal - is straight forward and you only need to follow the offical upgrade instructions. Short version on the console looks like this: sudo do-release-upgrade This will update the repository entries, and start the upgrade process. After some minutes or hours of download and installation, you have to reboot your system once to get the new kernel loaded. As time of writing, I'm on '3.5.0-17-generic'. And as with any modification of the kernel version, you have to compile the necessary kernel modules to get VMware Player or Workstation up and running. Usually, this happens the first time you try start your VMware software and that's it. Well, again not so this time. Getting the kernel patch Luckily, the community over VMware is very active and you can get a new kernel patch in the online forums here. Get the download and put in a folder have write permissions. Then you extract the archive on the console like so: tar -xjvf vmware9_kernel35_patch.tar.bz2 Then you change into the newly created folder: cd vmware9_kernel3.5_patch/ And you execute the available shell script as root (superuser) like so: sudo ./patch-modules_3.5.0.sh This will stop any running instances of VMware software, patches the source files and runs the compile process for your active environment. This might take some time depending on your machine, and once completed you can start VMware Player or Workstation as previously. In case that you are going to apply the patch again, the script will simply quit with the following output: /usr/lib/vmware/modules/source/.patched found. You have already patched your sources. Exiting You might remove the .patched file in case that you upgraded/changed your kernel and you need to apply the patch again. Disclaimer: The patch is "as-is" and the patcher is originally created by Artem S. Tashkinov, and later modified by An_tony. Please refer to the VMware forum in case of questions or problems. There are also patches available for older versions of VMware Player or Workstation.

    Read the article

  • What sort of things can cause a whole system to appear to hang for 100s-1000s of milliseconds?

    - by Ogapo
    I am working on a Windows game and while rendering, some computers will experience intermittent pauses ("hitches" for lack of a better term). When profiled they appear in seemingly random places in the code. Eventually I noticed that it wasn't just my process that was affected, but (seemingly) every process on the system. All of the threads in my application hitch at once. The CPU utilization drops during these hitches and it appears as if most processes make no progress. This leads me to believe this may be an Operating System or Driver issue, but it only occurs while playing the game (and only on some systems). What sort of operations might the operating system be doing that would require the kernel to pause all user threads and block. Some kind of I/O? At first I thought of paging but my impression is that would only affect a single process, no? Some systems in use: Windows, DirectX (3d), nVidia cards (unknown if replicates on ATI), using overlapped io for streaming

    Read the article

  • How to work with processes?

    - by Viesturs
    I have seen similar questions here, but I didn't get my answer. Maybe it's because I am new to all this and just don't understand. I want my app to work mostly as an indicator. And if user would start it again it would check if it is already running, if it is then give all the input data to that process and quit. So first I need to check if it is running. I saw the answer where you can make a file witch when the program starts and then check if it exists... But what if someone would delete it? Can't I just ask the OS if there is process named "myApp" or something? The next thing I don't really get is how to communicate with the process. How do I give it the input data and what is it going to do with it? Does it work just like starting a new app, through the main() method? I am trying to create this using Quickly. So it would be nice if you can give me some python examples or link to something like that.

    Read the article

  • Ubuntu 11.10 USB 3.0 HDD

    - by Chazm
    I have a problem with my external HDD (WD My Book Essential 1TB) working on USB 3.0 port. I'm using dual boot setup with Windows 7 and Ubuntu 11.10, both 64 bit. While I'm running Windows and rebooting back to Windows everything works well. When I'm switchin to ubuntu everything works great as well. But after 1st reboot from ubuntu neither windows nor ubuntu mount the external drive. I have to reboot the device manually. I suspect that the problem is with unmounting the device on shutdown on ubuntu. The case concern only usb 3.0. When i plug the same device into usb 2 port the problem doesn't persist. Does any1 hit the same problem or have a clue what might goin wrong?

    Read the article

  • Ubuntu audio mysteriously stopped working (12.04)

    - by Laika
    Well, I've been a user of Ubuntu 12.04 LTS since April now, and it's been a very pleasant experience. I'm a big fan of electronic music, and I tend to have my tracks playing in the background while I do things on my laptop, either in YouTube or in Clementine, my default music player. All has worked very well until now. A couple of days ago my entire PC started to lag really badly. Almost everything was unusable. I opened up System Monitor via the terminal to find a process called "pulseaudio" using nearly 1GB of RAM and over 80% of my CPU. I needed to get some important work done and so I killed the process without thinking. Once again today, pulseaudio decided to lag the hell out of my PC, and so I killed it again. Nothing seemed to happen immediately, but once I opened up YouTube all the audio on videos stuttered a lot, while the videos played smoothly. I restarted Firefox to find that the audio was now not working at all, with both headphones and speakers, and the volume up quite a bit (it's not muted, I've checked that!). A little bit of research later and I've discovered that pulseaudio plays an important part in Ubuntu's audio. Even after restarting my PC the audio still ceases to work in any applications or with any output. The pulseaudio process refuses to start up again. So, can you help me out here? What can I do to fix my problem, and why was pulseaudio doing this in the first place?

    Read the article

< Previous Page | 231 232 233 234 235 236 237 238 239 240 241 242  | Next Page >