Search Results

Search found 6200 results on 248 pages for 'lib'.

Page 173/248 | < Previous Page | 169 170 171 172 173 174 175 176 177 178 179 180  | Next Page >

  • Applications starts very slowly from a network path

    - by Snowfox
    Hi We have a windows 2008 server which hosts the network share \\srvcompany\lib. This share contains several applications needed for the daily business. Every client/user (all win xp) has shortcuts on the desktop to these apps. We have the problem that at several (but not all) clients the apps starts very slowly. If I copy the application's programm files to a local folder then they'll start fastly. When I watch the memory usage in the task manager on such a "slow" machine while an applications starts I notice that the memory usage grows much slowier than when I start the app from a "fast" machine. But when I copy files with Windows Explorer from this share, the speed is nearly the same. I've also checked the network driver, both tested clients have the same network card with the same driver version. Has anyone an idea where or what I should check next to solve this problem? Thanks for answers.

    Read the article

  • Ubuntu server 12.04.03 not checking disk partitions on reboot?

    - by jamesc
    My MOTD is showing: *** /dev/md2 will be checked for errors at next reboot *** *** /dev/md1 will be checked for errors at next reboot *** *** /dev/md3 will be checked for errors at next reboot *** However, a standard sudo shutdown -r now does not appear to check the disks and the message remains. My guess is that the partitions 'should' be checked and that they are not being... so how can I get Ubuntu to check the partitions at reboot and keep things nice and safe? Update - this is the output of cat /etc/fstab proc /proc proc defaults 0 0 none /dev/pts devpts gid=5,mode=620 0 0 /dev/md/0 none swap sw 0 0 /dev/md/1 /boot ext3 defaults 0 0 /dev/md/2 / ext4 defaults 0 0 /dev/md/3 /home ext4 defaults 0 0 Update 2 - One message gone... Using @christianwolff's suggestions... sudo rm /var/lib/update-notifier/fsck-at-reboot sudo touch /forcefsck sudo shutdown -r now And now the motd is down to *** /dev/md1 will be checked for errors at next reboot *** *** /dev/md3 will be checked for errors at next reboot *** So 'md2(ext4)' has been checked and the message updated.

    Read the article

  • How to migrate Lotus Notes Mail in database to Exchange public folder?

    - by elsni
    I need to migrate a Lotus Notes Mail database to Exchange. In the past, Users get mails in their Notes mail accounts and sort them manually by drag and drop in a specific folder structure in a seperate Notes mail database. This should be mapped to Exchange. I considered using outlook and a sharepoint Library mapped to outlook, but Outlook does not support drag and drop from the mail account to a standard doclib and the discussion libs do not support folders. So I think it's the easiest way to use an Exchange public folder instead of a sharepoint lib, which should work as in Notes (correct me if I'm wrong). But how I do migrate the old Notes db to the public folder including all subfolders? Thank you!

    Read the article

  • SSL issues with puppet agent at openSUSE

    - by Roman Grazhdan
    I have a master running at my vps, and it has a simple helloworld manifest which works fine with any ubuntu machine I have. It connects, exchanges keys and creates test file allright, so I'm sure it's not server issue. The agent which is running at a virtual machine with openSUSE says: err: Could not request certificate: SSL_connect returned=1 errno=0 state=SSLv3 read server certificate B: certificate verify failed. This is often because the time is out of sync on the server or client I believe it's probably a broken or missing lib, since the package is not built very accurately - it wouldn't start out of the box because of wrong path to lockfile, for example. So how do I figure out what exactly is wrong here? The time is allright, I've checked it. I probably could do without SSL if it's possible, since that SUSE machines are just for training, but it's the last opportunity.

    Read the article

  • Building modules with linux kernel for custom flavor

    - by Chandan Apsangi
    I followed the instructions given in the link: http://blog.avirtualhome.com/how-to-compile-a-new-ubuntu-11-04-natty-kernel/ for building a custom kernel and booting it. Everything works fine, except that when building it, I used the option skipmodule=true (as given in this link), so I guess the modules are not built for this kernel. I'm saying this because after I installed the image (using dpkg -i), when I run 'ls -l' in the /lib/modules/2.6.38-11-i5 folder, all the files/folders except the 'kernel' folder have got updated (Btw, I had built it last time and it had worked fine,this is an update on the same kernel source). So is there some option/command that I missed that exports the modules as well? Thanks for your help.

    Read the article

  • How to specify which GCC for MacPorts to use?

    - by penyuan
    I compiled GCC 4.4.3 and installed it in /usr/local/bin, but whenever I install a port via MacPorts 1.8.2 the verbose output says MacPorts is using /usr/bin/gcc-4.2: checking for gcc... /usr/bin/gcc-4.2 How do I make MacPorts find my own GCC 4.4.3? Here is my existing path: /opt/local/bin:/opt/local/sbin:/usr/local/bin:/usr/local/lib:/usr/bin:/bin:/usr/sbin:/sbin:/usr/local/bin:/usr/X11/bin Thank you! P.S. I am running Mac OS X 10.6.2 Snow Leopard.

    Read the article

  • How to allow Hudson build URL through Nginx auth_basic?

    - by rodreegez
    Hi, I have Hudson running and made available to the world via nginx. I have protected Hudson with nginx's auth_basic and that works great. The trouble is, I want to allow unauthenticated requests to the build URL, i.e. /job/<job_name>/build. Currently I have this in my nginx conf: upstream hudson { server 127.0.0.1:8888; } server { server_name ci.myurl.com; root /var/lib/hudson; location / { proxy_pass http://hudson/; auth_basic "Super secret stuff"; auth_basic_user_file /var/opt/hudson/htpasswd; } location ~ \/build { auth_basic off; } } I can't get that second location to allow unauthenticated requests. I have tried various combinations of location ~ /job/(.*)/biuld { } location ^~ \/build { } location ~ \/job\/(.*)\/build { } etc... Maddening! Can anyone point me in the right direction? Thanks, Ad.

    Read the article

  • How to limit server to specific IP addresses with mod_authz_host?

    - by BeeDog
    Hi! I am very new to this area, so please bear with me. :) Right now I am running an Apache HTTP server on my setup, a very basic configuration. The website hosted on it is accessible from anywhere, and I want to limit the access to a specific IP address range. I've looked into this and I found that one Apache module called mod_authz_host handles this. http://httpd.apache.org/docs/2.2/mod/mod_authz_host.html The problem is, I haven't managed to find documentation that explains well how to actually do the stuff. How do I actually make sure only a certain range of IP addresses can access my site/server? The machine is running Ubuntu Server 10.10, the web files are stored in /var/www/, the apache2 daemon has its stuff stored in /etc/apache2/ and /usr/lib/apache2/modules/*. Thanks in advance, and sorry if this is a stupid question!

    Read the article

  • MySQL Optimizing

    - by Thoman
    Hello My web use an dedicated Intel(R) Xeon(R) CPU E5620 8core 12Gram Centos32bit/Driectadmin DISK SAS 80G Php-cgi This dedicated running one website Use wordpress 2.92(+plugin cache...) Database size 600MB only 100online But mywebsite runing very snow. please hep me config file my.cnf [mysqld] user=mysql key_buffer=128M set-variable = max_connections=1000 socket = /var/lib/mysql/mysql.sock key_buffer =32M table_cache = 1024 open_files_limit = 16344 join_buffer_size = 8M read_buffer_size = 8M sort_buffer_size = 8M tmp_table_size=512M read_rnd_buffer_size=8M max_heap_table_size=256M #myisam_sort_buffer_size=256M thread_cache_size=8 thread_cache=32 query_cache_type=1 query_cache_limit=1024M query_cache_size=1024M thread_concurrency = 16 wait_timeout = 10 connect_timeout = 10 interactive_timeout = 10 long_query_time=1 log-slow-queries = /var/log/mysqlslowqueries.log max_allowed_packet=32M skip-innodb [myisamchk] key_buffer = 64M sort_buffer = 64M read_buffer = 16M write_buffer = 16M [isamchk] key_buffer=64M sort_buffer=64M read_buffer=16M write_buffer=16M And apache

    Read the article

  • Python - How to remove/unimport libs was imported before

    - by Marslo
    As we know that, in python 2.x, integer would be got if we divide two integer values. However, if using the furture (it's might be a lib or something like that), just like from __future__ import division, we can get float value. E.g.: >>> 3/2 1 >>> from __future__ import division >>> 3/2 1.5 >>> >>> >>> 3//2 1 >>> 4/3 1.3333333333333333 >>> So, '//' instead of '/' should be used if getting integer after imported division, but I want to know how to using '/' to get integer again. That is mean, whether there is some way to un-import or remove the libs which was imported before.

    Read the article

  • What's the best way to install mod-wsgi for a specific version of python (2.7) on Debian (squeeze)?

    - by Pascal Polleunus
    System info: Debian Squeeze Python versions: 2.6.6 (default) & 2.7.2 mod-wsgi: 3.3-2 in /usr/lib/apache2/modules: mod_wsgi.so -> mod_wsgi.so-2.6 mod_wsgi.so-2.5 mod_wsgi.so-2.6 I've a virtualenv configured with Python 2.7 but my application runs 2.6.6, apparently because mod-wsgi is using the default version of the system. The way to go seems to configure mod-wsgi to use 2.7 (i.e. system-wide, not specific to a virtualenv). How can I install mod_wsgi.so-2.7? After installing that, I'll just have to change the symlink mod_wsgi.so to mod_wsgi.so-2.7.

    Read the article

  • MPICH2 vs KERRYGHED

    - by user40135
    Hi All right now I am moving first steps in clustering. I installed MPICH2 on my Ubuntu at home and I have a silly question about it. For what I am reading right now it seems that it provides the capability of sending processes to other pcs. I went for this lib just because I set it up very quickly and easily. Compared to MPICH2 , do you know what is the advantage of having a different clustering system like KERRYGHED? It seems that these ones also provide this capability, but the Kernel must be rebuild, so I suppose that it is going to be faster. What other advantages are remarkable for a clustering system like this? Thanks

    Read the article

  • Apache has many PHP session files

    - by PiTheNumber
    # ls /var/lib/php5 | wc -l 7488 # ls -la -rw------- 1 wwwrun www 0 Nov 9 15:30 sess_vtuh671rlafdidfjmgjfu6065p4tfieg -rw------- 1 wwwrun www 0 Nov 12 02:30 sess_vu9pn476oiqbsd20q4s2brt60b9vg90d -rw------- 1 wwwrun www 0 Nov 9 15:07 sess_vuonfs2cqsdiq8ja51ornh6lp5j9mf93 -rw------- 1 wwwrun www 0 Nov 9 16:02 sess_vuutcad8as55il34db3uqhqrsltd4q6o -rw------- 1 wwwrun www 0 Nov 9 23:26 sess_vv2mrv5dnlnts6das4g5jlfldael4l0e -rw------- 1 wwwrun www 44 Nov 9 20:35 sess_vvc0cfjuvk3lqb5m97fv6gsmv6bjhsdk -rw------- 1 wwwrun www 0 Nov 9 10:33 sess_vvq82fhj9lg29gaejemlb2lrk25mqv7d -rw------- 1 wwwrun www 0 Nov 9 20:36 sess_vvtd4ka8rfmcroa34unl06916ubj8sb9 Most of them are empty. There are not so many users on the server so I wonder where those files came from. Is this a problem or how does apache handle those file? Do they get delete automaticly? Could this be caused by a bad PHP file?

    Read the article

  • Firewall blocks outgoing email

    - by Martin Trigaux
    On my Debian server running a Django website, I have an error when I need to send an email. The error received is Exception Type: gaierror Exception Value: [Errno -2] Name or service not known Exception Location: /usr/lib/python2.6/socket.py in create_connection, line 547 You can see the full error log here. After testing, it seems it is my firewall that blocks the request. You can see my iptable file (/etc/init.d/firewall). I think the problem comes from the two commented lines that were supposed to accepts all established connections. When I uncomment them, I have an error iptables: No chain/target/match by that name. Thank you

    Read the article

  • In Linux, is it possible to get a listing of drives' disk space usage that also shows volume labels?

    - by DavidH
    I know about df, of course, but df does not output volume labels. I have 5 USB hard drives plugged into my NAS box, and would love to know which is which. Current df output: Filesystem Size Used Avail Use% Mounted on /dev/sda1 27G 2.2G 24G 9% / none 56M 476K 55M 1% /dev none 60M 0 60M 0% /dev/shm none 60M 332K 59M 1% /var/run none 60M 0 60M 0% /var/lock none 60M 0 60M 0% /lib/init/rw /dev/sde1 150G 102G 48G 68% /media/usb0 /dev/sdb1 299G 196G 103G 66% /media/usb1 /dev/sdc1 233G 183G 51G 79% /media/usb2 /dev/sdd1 233G 209G 25G 90% /media/usb3 /dev/sdf1 150G 101G 49G 68% /media/usb4

    Read the article

  • why is this happening?-"dhcpcd will not work correctly unless run as root"

    - by user330317
    i have installed archlinux and gnome on virtualbox. had no problem connecting to internet but now after installing gnome and rebooting there is no internet connection after following instructions from archwiki,i have tried . but i cant figure out the problem please help. host-63drhd% sudo netctl status enp0s3 ? [email protected] - Networking for netctl profile enp0s3 Loaded: loaded (/usr/lib/systemd/system/[email protected]; static) Active: inactive (dead) Docs: man:netctl.profile(5) host-63drhd% sudo netctl enable enp0s3 Profile 'enp0s3' does not exist or is not readable host-63drhd% sudo dhcpcd dhcpcd[1486]: sending commands to master dhcpcd process host-63drhd% dhcpcd dhcpcd[1543]: control_open: Permission denied dhcpcd[1543]: dhcpcd will not work correctly unless run as root dhcpcd[1543]: open `/run/dhcpcd.pid': Permission denied dhcpcd[1543]: control_start: Permission denied dhcpcd[1543]: version 6.3.2 starting dhcpcd[1543]: enp0s3: if_init: Permission denied dhcpcd[1543]: enp0s8: if_init: Permission denied dhcpcd[1543]: no valid interfaces found dhcpcd[1543]: no interfaces have a carrier dhcpcd[1543]: forked to background, child pid 1544

    Read the article

  • MySql transfer / update (a bit specific)

    - by Jeff
    before posting I was digging whole site but didn't find help for my problem, so I hope someone will help... Facts: 30 Gb mysql database on remote server (about 20.000.000 rows) data are once weekly updated in local network (mysql) I need to transfer/replace local updated database with remote connection is about 2mb (real mb, not mbps) up/down Point is that I can't have 'down time' of remote mysql server. Until now I Tried: navicat data sync - Ok, but take about 3 days to finish dbForge - ok but need 5 days to finish mysql dump transfer to remote server and execution - about day, but a lot of downtime rsync folder with database /mysql/lib/MY_DATABASE - 4 hours, but after that I need to execute always 'repir on remote server' which takes about 2 hours, and a lot of down time mysql dump piped from cl to directly goto server - still now satisfied many problems I could give you more things that I tried... mysql replication - slow Anyase, what is best,best way to: refresh remote mysql on weekly level and in same time to have 0 sec down time nor huge server load If you have any idea please share

    Read the article

  • Windows/Samba connection error

    - by Gomibushi
    I have a Linux fileserver serving up /home for linux and windows users. I was able to connect from my windows client, but not from a DC. Then suddenly I could connect from the DC too. The linux servers run Centrify clients, and as such are part of the domain. All on same subnet. This is what the the log.smbd says, repeatedly: [2010/02/11 11:25:57, 0] lib/util_sock.c:read_data(534) read_data: read failure for 4 bytes to client 192.168.200.3. Error = Connection reset by peer On Windows it appeared as an "unknown error". EDIT: the error code is "0x80004005". We are developing a system depended on the samba share, and are worried this will appear again. It would be nice to pin point the root of this. Any ideas what this might be? Places to look?

    Read the article

  • Postgres Remote Access

    - by boot-baby-boot
    I am trying to connect to postgres remotely.I have followed this tutorial http://www.cyberciti.biz/faq/howto-fedora-linux-install-postgresql-server/ and have executed the following commands to see if the remote access is possible. [root@printmyworld ~]# egrep -i "(listen_addresses|port|tcpip_socket).*=.+" /var /lib/pgsql/data/postgresql.conf #listen_addresses = '*' # what IP address(es) to listen on; #port = 5432 [root@printmyworld ~]# lsof +c0 -anPiTCP -upostgres COMMAND PID USER FD TYPE DEVICE SIZE NODE NAME postmaster 9323 postgres 3u IPv4 2875987353 TCP 127.0.0.1:5432 (LISTEN ) postmaster 9323 postgres 4u IPv6 2875987354 TCP [::1]:5432 (LISTEN) I am suspicious of this line: postmaster 9323 postgres 3u IPv4 2875987353 TCP 127.0.0.1:5432 (LISTEN My server ip address is 1yy.000.1xx.000 .Should it be 1yy.000.1xx.000:5432

    Read the article

  • How to arrange 2 SSD with 2 SATA?

    - by alfish
    I like to have best io performance as well as good capaciyy and reliability out of a server that hosts a busy forum, which involves loads of static files download. I am wondering what is the best plan to format and use the disks given that the server has only 4 disk bays and I have 2 SSD and 2 SATA disks at hand. I am currently thinking about putting the disks in RAID 10 so that SSD contains /var/lib/mysql as well as most of the OS (Likely to be Debian) and SATA disk to contain /path/to/static/files. However I'd like to hear your expert opinion on this. Thanks

    Read the article

  • can /usr/src be a sym link?

    - by lord.didger
    I want to store all source code of programs I have installed in /usr/src. However, due to size of the drive I made a sym link /usr/src that points to ~/src. That was nice. Unfortunately, that caused virtualbox-dkms to fail to build the virtualbox kernel module because of a symlink within the linux-headers-*-common. 'script' points to ./../lib/linux-kbuild-3.1/scripts what is fine in /usr/src directory but wrong in ~/src. Can I bypass this problem or the only solution is to store sources within the directory /usr/src?

    Read the article

  • Web-Server directory permissions

    - by MLS
    Hello All, I would like some help understanding web-server directory permissions. Apache, CentOS, PHP, Mysql Example, I have multiple sites in /var/www/html They are in paths like: /var/www/html/www_domainname_com inside each site I might have a path like /lib/mysql/ like PHP connect stuff, database config, etc. What should me permissions be so that someone cannot just browse to that directory? Should I just .htaccess them? I have apache:apache as the owner of all my web directories. Can I prevent someone from crawling certain directories of my web-server? I have a robots.txt, but what is to say the crawler obeys it? So to sum up: 1. What is the best owner/permission set for my sensitive files that the web-server or php or mysql needs, but I dont want people browsing to? Can I prevent straight out crawling of portions?

    Read the article

  • Where are incoming messages queued/stored when sendmail is communicating with a milter?

    - by Mike B
    CentOS 5.x | SendMail I'd like to better understand how milter hand-offs work. We a remote MTA initiates an SMTP session, my understanding is that Sendmail hands the message data off to the milter during that session. Where/how is this message data stored? Is this all done in memory? Or is there a "queue" area where message data for milter is stored for pickup by the milter? If it helps, let's say I'm calling a milter like this: INPUT_MAIL_FILTER(`foo', `S=unix:/var/lib/foo/foo.sock, F=T, T=C:5m;S:3m;R:5m;E:5m')dnl

    Read the article

  • linux system problem

    - by snakec
    very first thanks to u all for Ur support. I'm quite new to Linux .i know how to install software but i don't know 1: how to install library .a or .so files i: how to install tar.gz i use the method like ./configure. make make install but most of the time i got the message nothing to make .in lots of tar.gz there is no installation document no make file no .configure file that make me quite confused how to install them or run them now i got sample source code of cuda i got them in tar.gz form when i extract them i found a folder in folder i found folder like c ,doc,shared etc when i open each folder i found more folder n file like that src, doc common ,lib, in these folder i found source code file header files libraries file make files i don't know how to run this kind of project can the be installed on the system how to run them they don't have .run file or script they don't have configure file can any one explain me how to compile them ,how to run them & how to install them

    Read the article

  • Deploy multiple django instances on one Host [migrated]

    - by tvn
    I am trying to setup multiple Django instances on one Host with lighttpd. My problem is to get Djangos FCGI working on subdirectories served by my Webserver. So my aim is the following: www.myhost.org/django0 - django1.fcgi on localhost:3000 www.myhost.org/django1 - django2.fcgi on localhost:3001 www.myhost.org/django2 - django3.fcgi on localhost:3002 Unfortunately the following configuration doesn't even work for one: $HTTP["url"] =~ "^/django0/static($|/)" { server.document-root = "/home/django0/django/static/" } $HTTP["url"] =~ "^/django0/media($|/)" { server.document-root = "/usr/lib/python2.7/dist-packages/django/contrib/admin/media/" } $HTTP["url"] =~ "^/django0($|/)" { proxy.server = ( "" => ( ( "host" => "127.0.0.1", "port" => "3001", "check-local" => "disable", ) ) ) } The only response I get is an 404 and even this takes a long time till I get this. I found nothing suspicious neither in the access.log nor in the error.log.

    Read the article

< Previous Page | 169 170 171 172 173 174 175 176 177 178 179 180  | Next Page >