Search Results

Search found 55789 results on 2232 pages for 'ubuntu 13 10'.

Page 418/2232 | < Previous Page | 414 415 416 417 418 419 420 421 422 423 424 425  | Next Page >

  • Roundcube "Server Error (OK!)": Lists no messages but can get messages according to the log file

    - by thonixx
    In my server setup there are three virtual machines. One windows machine, an Ubuntu Server 11.10 and a Debian Squeeze mailserver. On the Ubuntu system I have Roundcube installed and I want to connect to the virtual mail server. What's the problem After login into Roundcube it says "Server Error (OK!)" and lists no messages. More information On the Ubuntu server there is no error in any log file (even Roundcubes log files). In the imap log file there you can see Roundcube is able to fetch all imap messages (I can see them in the imap log file created by Roundcube). And on the side of the mail server there are no error messages too. The test connection at the end of the configuration of Roundcube works too, there is a "success" notification. Even the basic login at Roundcube login dialog works without any error message. Roundcube log file you can look here for the log file: http://fixee.org/paste/wxg36eh/ So does anyone know what's wrong with Roundcube?

    Read the article

  • Is is possible to hide label colors in list view in Mac OS 10.X Finder? I use labels as star ratings

    - by Andrew Swift
    I have modified the first five label names in OS X Lion to be ? through ?????. This way I can easily tag photos etc. in the finder. However, I find that Apple's rendering of the colored bars is horrendously ugly. I'd like to keep the labels (to be able to sort by label) but not see the colors. I know it was possible to change the colors with Label X, but Unsanity has not issued an update for Lion. I don't need to change the colors, just find a way to hide them in the finder.

    Read the article

  • Whats the best cloud backup solution for a small scale server envoirnment?

    - by nbv4
    I have a server that runs a postgres database that contains about 200MB of data. Currently I have a cron job setup on my home computer which: ssh's into my server runs a remote script which makes a backup of the database scp's that dump over to my local hard drive for storage. Each dump is 20MiB. does this every six hours (one months of backups is roughly 2GiB) The problem with this setup is that if my local machine goes down for whatever reason, no backups will be made. Also, I can't have the cron run from the server, because I can't have it scp'd to my local machine from my server (firewalls and all that crap). My local machine is running Ubuntu 10.04, and my server is Ubuntu 9.10 server edition. I looked into Ubuntu One, but currently it's gui-only. I also looked into dropbox, but it's a pain in the ass to get setup in linux without gui support. Amazon S3 looks good but it's not free (yet dirt cheap). Is there any other alternative that I should look into? I'd prefer something where I can just have my script dump the database into a directory, and have the backup service 'watch' that folder and sync accordingly. I can maybe also have my local machine sync to the cloud backup so I have even more redundancy, plus easy access to my backups for use in testing.

    Read the article

  • Accesing internet with Host-Only network VirtualBox

    - by Bibo
    I have two VMs in my VirtualBox. Both are FreeBSD and should communicate between each other. I have created network with DHCP enabled and set like this: Server Address 10.10.10.1 Server Mask 255.255.255.0 Lower Address Bound 10.10.10.100 Upper Address Bound 10.10.10.200 My VMs are pinging each other and getting right ip adress. This is what I must have and can´t change. Now I want to access internet from one VM and I found that I must use two network cards but I don´t know how to set the second one. I am new with VirtualBox and this stuff so if you count with this in you answer I would appreciate that. And I have Windows 8 as host system if it´s matter.

    Read the article

  • Print over remote CUPS server, but just show a subset of the printers.

    - by jdm
    I'd like to print from my Ubuntu laptop (karmic) to some networked printers. Our organisation uses a CUPS server with several hundred printers. What I know I can do is: CUPS_SERVER=printers.company.com acroread document.pdf and then Adobe Reader shows me all available printers to select from. However, it takes a couple of minutes to display the large list, which is really annoying. (The desktop PCs here suffer from this, too.) The other option is to add a new printer with an address like ipp://printers.company.com/printer/bldg1_hp8150 (to the Ubuntu printer configuration = local CUPS server). However, it asks me for a driver. I don't want to / can't always specify a driver, since some printers don't appear in the list. I'd like to let the remote CUPS server handle the driver part (like it does when i set CUPS_SERVER), and do no more preprocessing/"driver stuff" on my side. The ideal thing would be if I could somehow add the remote printer list to my local cups server, and apply a filter, so that it would just display printers a la bldg1_*. This feature was available in KDE3.?, but I can't find something similar in Ubuntu/Gnome. Any suggestions?

    Read the article

  • Apache local verses external (domain)

    - by Jessy Houle
    I have an Apache server running on Ubuntu server 10, using Passenger for Ruby on Rails. I have configured my site under the sites-enabled directory of Apache and can hit the server with an internal IP address (192.168.X.X) and the site comes back as expected. However, whenever I try to hit the site externally, either through the domain name or the IP address tied to the domain name, the site will not come back. I have a router in the middle with a static IP address, with Port Forwarding turned on (forwarding 80/443) to the server and I'm quite confident the issue isn't there. In fact, I even DMZed to the Ubuntu Server just to make sure. Also, all router firewall options have been turned off. So here is the question... Is there something else I have to do with Ubuntu server to allow externally requested port 80 traffic? Otherwise, is there some settings that need to be set in Apache to allow domain or external IP address port 80 traffic through? I'm pretty new to Apache, so, please take it a bit easy on me :-) Thank you for your responses. -Jessy Houle

    Read the article

  • Whats the best cloud backup solution for a small scale server environment?

    - by nbv4
    I have a server that runs a postgres database that contains about 200MB of data. Currently I have a cron job setup on my home computer which: ssh's into my server runs a remote script which makes a backup of the database scp's that dump over to my local hard drive for storage. Each dump is 20MiB. does this every six hours (one months of backups is roughly 2GiB) The problem with this setup is that if my local machine goes down for whatever reason, no backups will be made. Also, I can't have the cron run from the server, because I can't have it scp'd to my local machine from my server (firewalls and all that crap). My local machine is running Ubuntu 10.04, and my server is Ubuntu 9.10 server edition. I looked into Ubuntu One, but currently it's gui-only. I also looked into dropbox, but it's a pain in the ass to get setup in linux without gui support. Amazon S3 looks good but it's not free (yet dirt cheap). Is there any other alternative that I should look into? I'd prefer something where I can just have my script dump the database into a directory, and have the backup service 'watch' that folder and sync accordingly. I can maybe also have my local machine sync to the cloud backup so I have even more redundancy, plus easy access to my backups for use in testing. Edit: My server is a VPS, so what solution I end up using has to be 100% software based.

    Read the article

  • Windows 2003 - RAS service - VPN client can only connect to server, not internal network

    - by Gk
    Here is my diagram http://imageshack.us/photo/my-images/263/vpnt.png/ Server has two NIC: One connect to Internet, one connect to internal network. RAS service is configured with static IP pool on the same subnet of internal network (10.10.0.x). The problem is when a client connect to server, client can connect to server(10.10.0.10), but cannot connect to another host locate in internal network (10.10.0.11) even cannot ping to 10.10.0.50. RAS configured with IP routing enabled, client connect with options: Use default gateway on remote network. I'm messed up and can not figure out what wrong. Please help me T_T. TIA, giobuon

    Read the article

  • Some of my keys are automatically being pressed along with other keys

    - by Santosh
    History The last time when my computer shutdown was a power failure. Now some keys are automatically being pressed when I type something. The last thing I did to keyboard setting was adding a keyboard layout (on Ubuntu). What is happening Whenever I press c, xc is writeen s gives me sd d gives me sd e gives me we 2 gives me 23, So when I want @ it gives me @# 3 gives me 23 Pressing CAPS Lock gives me F3 and vice-versa. All other key are either working fine or I don't use them. I have two operating system Ubuntu and Windows, I use Windows very less and found this problem on Ubuntu, but as soon as I logged in to Windows (for checking) then I found that Windows has the same problem. Effects on my life This starts form the time of login, even I have problem in typing my password. Whenever I try to save any webpage, it is bookmarked automatically. Whenever I copy, it is cut automatically. I have to spend more than half of time correcting what I have typed. Note: Typing thisd quwesdtion wasd rweally a big pain to mwe.

    Read the article

  • /etc/hosts: What is loghost? (fresh install of Solaris 10 update 9)

    - by cjavapro
    # # Internet host table # ::1 localhost 127.0.0.1 localhost XX.XX.XX.XX myserver loghost What is the purpose of loghost? If it was not for having loghost in there, all the /etc/hosts files on all the servers in this particular network could be identical. Edit: I looked at /etc/syslog.conf #ident "@(#)syslog.conf 1.5 98/12/14 SMI" /* SunOS 5.0 */ # # Copyright (c) 1991-1998 by Sun Microsystems, Inc. # All rights reserved. # # syslog configuration file. # # This file is processed by m4 so be careful to quote (`') names # that match m4 reserved words. Also, within ifdef's, arguments # containing commas must be quoted. # *.err;kern.notice;auth.notice /dev/sysmsg *.err;kern.debug;daemon.notice;mail.crit /var/adm/messages *.alert;kern.err;daemon.err operator *.alert root *.emerg * # if a non-loghost machine chooses to have authentication messages # sent to the loghost machine, un-comment out the following line: #auth.notice ifdef(`LOGHOST', /var/log/authlog, @loghost) mail.debug ifdef(`LOGHOST', /var/log/syslog, @loghost) # # non-loghost machines will use the following lines to cause "user" # log messages to be logged locally. # ifdef(`LOGHOST', , user.err /dev/sysmsg user.err /var/adm/messages user.alert `root, operator' user.emerg * ) Very interesting. when shutting down,, alerts go to all users probably through *.emerg * Looking at ifdef, it seems that the first parameter checks to see if current machine is a loghost, second parameter is what to do if it is and third parameter is what to do if it is not. Edit: If you want to test a logging rule you can use svcadm restart system-log to restart the logging service and then logger -p notice "test" to send a test log message where notice can be replaced with any type such as user.err, auth.notice, etc.

    Read the article

  • Solutions on how to use an OS X calendar as a more perfect time tracking solution for 5-10 users in a small agency?

    - by jnthnclrk
    I really like OS X's iCal. Entering events is easy with the mouse and it also gives you a very real visual sense of how long tasks take to complete. We often work remotely in our organisation, so we use a few shared calendars between key individuals to provide us with an overview of hours worked, availability & schedule conflicts without too much disruption to our various, hectic workflows. It really is a neat solution, especially on shared tasks. How many times have you tasked a remote colleague and then lost the thread on whether that task was completed or not? With shared calendars you get a much clearer idea of what your people are working on without having to pick up the phone or compose a chat. However, there are a few areas where this approach fails... iCloud syncing often needs to be re-jiggered The "view only" option on shared calendars does not seem to work, which makes all shared calendars editable by others There is no decent reporting with this workflow There is no task categorisation or tagging Things get very busy in iCal when working with more than 2 shared calendars I've looked at a few task management apps like Basecamp and Harvest, but nothing appears to let me edit my calendar natively and then sync with a 3rd party. Interested in solutions to improve the above workflow and enable us to elegantly increase the amount of users.

    Read the article

  • Active Directory password update not recognized on OS X 10.7.3?

    - by eaj
    We have a few Macs in our office which are bound to our Active Directory (Server 2k8R2) domain for login authentication and access to file shares. For the most part this has worked without trouble. I recently changed my AD password (from the Windows box that is my main desktop machine), and my iMac still wants my old password for logins—even after a reboot. It correctly prompts for the new password when I attempt to connect to Windows smb file shares. The last time I changed my password, the iMac was running Snow Leopard and picked up the password change as expected. I'm a little bit of an OS X n00b, so I apologize if this should be obvious. My searching has turned up numerous issues people have had with Lion and Active Directory, but not this particular problem. Possibly related, we do have mobile accounts enabled for AD in the Directory Utility options. I assume this causes credentials to be cached locally, but it's not clear to me how to get them to update.

    Read the article

  • Computer not finding hard drives on boot -sometimes-

    - by todd.pund
    Computer specs: Mobo: Gigabyte ultradurable 3 - GA-970A-UD3 Processor: First gen I7 3.2GHZ Ram: 8GB Kingston DDR3 1066 Video Card: EVGA NVidia GTX 460 1GB Hard Drive: 500MB 7200rpm x2 (can't remember brand, sorry I'm at work.) Last week my developer preview for Windows 8 ran out so I put my copy of windows 7 back on the computer. The computer at that point started suffering from frequent freezing and crashing. When I rebooted the computer sometimes it wouldn't find the system HD at all. When I looked at the post screen it seemed to show that it wasn't finding either of the HDs. Then yesterday when turning on the computer I just got GRUB as a message (not a GRUB prompt, just GRUB) I haven't had a dual boot of Linux for at least a year. I loaded windows 7 recovery console from the disk and ran: bootrec /fixboot bootrec /fixmbr Which did not help. At that point I just installed Ubuntu 13.04 over the windows 7 install and still received the GRUB post. I went into the BIOS and switched the Hard Drive priorities and then it loaded into Ubuntu fine. For several days everything was just hunky dory until I installed the Ubuntu version of Steam, install Portal and tried to run it. At that point the computer froze and after hard rebooting couldn't find the hard disks again. Then after restarting the system it loaded up fine again and no issues since. (I have not tried to launch portal again). My next thought is to remove the system hard drive and try to use the secondary as the master to see if the primary HD is bad. I'm sorry if this has been confusing, I'll answer any questions I can. Any thoughts?

    Read the article

  • Why do I get swap space related errors when I still have lots of free memory in Solaris 10?

    - by Tom Duckering
    I am seeing a few of my services suffering/crashing with errors along the lines of "Error allocating memory" or "Can't create new process" etc. I'm slightly confused by this since logs show that at the time the system has lots of free memory (around 26GB in one case) of memory available and is not particularly stressed in any other way. After noting a JVM crash with similar error with the added query of "Out of swap space?" it made me dig a little deeper. It turns out that someone has configured our zone with a 2GB swap file. Our zone doesn't have capped memory and currently has access to as much of the 128GB of the RAM as it need. Our SAs are planning to cap this at 32GB when they get the chance. My current thinking is that whilst there is memory aplenty for the OS to allocate, the swap space seems grossly undersized (based on other answers here). It seems as though Solaris is wanting to make sure there's enough swap space in case things have to swap out (i.e. it's reserving the swap space). Is this thinking right or is there some other reason that I get memory allocation errors with this large amount of memory free and seemingly undersized swap space?

    Read the article

  • how can 192.168.2.10 talk to 172.18.156.65 ?

    - by steve
    from what i understand, all computers need to be in the same subnet /24 so how would i get a computer to be able to connect to another computer that is set as 172 ? one computer is behind a firewall, the other is connected to a HUGHES sat dish and has the IP of 172 would I need to setup a ROUTE in a firewall between the 2 computers, to say "if you are trying to get to 172.18.156.65 then use the ip address 192.168.2.65 and translate it to 172.18.156.65" the firewall is a netgear, if I should use a ROUTE, then I'll try to learn it from the manual thanks

    Read the article

  • Ways of file copy

    - by Tim
    I sometimes found that when using simple right-click and copy-and-paste, some files/directories are not copied completely or not at all, because of various reasons, such as some saved webpage files/directories have some strange characters in their names or their names are too long. For example, in Windows 7, I save this webpage http://www.howtogeek.com/howto/windows-vista/working-around-windows-vistas-shrink-volume-inadequacy-problems/ completely in a very deep directories whose parent directories may have long names, I cannot copy its top ancestry directory, as Windows complains the filename for the saved webpage directory is too long. In Ubuntu, sometimes I can save a file with some special character such as newline under some directory. But when I copy that directory, it will say the file name has some special character and I will have to manually remove the character. Such cases are complained in both Windows and Ubuntu. I was wondering what some better ways to accomplish the copy job in both Windows and Ubuntu. For example, will archiving all to be copied into a single archive help? If yes how to do that? Thanks and regards!

    Read the article

  • have 2 exchange servers to communicate together!

    - by Data-Base
    we have our exchange server 2007 (10.10.2.11) at work with domain "dddddd.com" and ironport and it work just fine! we need to test some thing at work with Exchange 2010 !! so we came up with this idea: we created a locked network and install firewall (10.10.2.88) the locked network has it's own DC and Exchange servers! we used the domain "dddddd.loc" any way it will be deleted! IPs DC 10.0.0.1 EX 10.0.0.11 now the Exchange server up and running in the firewall (10.10.2.88) we opened the ports and forwarded it to the Exchange2010 (10.0.0.11) in our main DC we defined the zone dddddd.loc to the firewall and the MX records as well but when we test we get error now how to make it so that the 2010 will send email directly to the 2007 (we will not use it to more that that)? and the 2007 will send the email to 2010 if the email has the domain dddddd.loc ?

    Read the article

  • Would it be better to have 10 GB PC5300 or 8 GB PC PC6400?

    - by jayrdub
    My development machine is heavy on memory usage. It currently only has 4GB of PC5300. I'm going to buy 8 GB of RAM, but should I take out all my existing memory and install just PC6400 or leave some of the original and stick with 5300? I don't know if I'll ever need more than 8GB, I might, I'm just wondering if the speed difference between 5300 and 6400 is worth trashing the 2 or 3 GB of old RAM I would still have room for after adding the new RAM.

    Read the article

  • Toshiba Satellite Laptop Realtek Network Driver Error code (10)?

    - by Silent
    hello all i have a Toshiba 64 bit windows vista, i tried looking for the drivers from the main website but no luck all the ones i got did not work or did not do anything. would someone know where i can find it or maybe fix this issue it will not find any network wireless or even if i hook up the Ethernet cable. i don't know why Toshiba makes it hard they should have a complete bundle. Model: A305-S687

    Read the article

  • Would it be better to have 10 GB PC5300 or 8 GB PC6400?

    - by jayrdub
    My development machine is heavy on memory usage. It currently only has 4GB of PC5300. I'm going to buy 8 GB of RAM, but should I take out all my existing memory and install just PC6400 or leave some of the original and stick with 5300? I don't know if I'll ever need more than 8GB, I might, I'm just wondering if the speed difference between 5300 and 6400 is worth trashing the 2 or 3 GB of old RAM I would still have room for after adding the new RAM.

    Read the article

  • Windows Broken after Deletion of HP_TOOLS partition

    - by beanland
    When I decided to install Ubuntu on my laptop as a dual-boot, I recognized that Windows 7 was using four separate partitions. I (yes, this was stupid) thought the HP_TOOLS partition was probably one I could get rid of, so I deleted it and installed Ubuntu side-by-side Windows using the installation wizard, but now Windows won't progress past the loading screen without the computer automatically restarting. I've had to use Ubuntu exclusively since then. I'm not sure how I can recover it. All of my files seem to still be there--I can mount the other partition and see them, use them, etc--but Windows 7 won't boot. I really have no idea what to do or what to try, or even if I'm at a salvageable point. here's a screenshot of GParted: This makes me suspicious that it wasn't necessarily the removal of my HP_TOOLS partition but the "unknown" status of that 992.50 KiB partition there, sda1. I'm assuming that's the recovery one? How can I get Windows bootable again? I'm sorry, but I'm so unfamiliar with this sort of thing I'm not quite sure where to start.

    Read the article

  • Would it be better to have 10 GB PC5300 or 8 GB PC6400?

    - by jayrdub
    My development machine is heavy on memory usage. It currently only has 4GB of PC5300. I'm going to buy 8 GB of RAM, but should I take out all my existing memory and install just PC6400 or leave some of the original and stick with 5300? I don't know if I'll ever need more than 8GB, I might, I'm just wondering if the speed difference between 5300 and 6400 is worth trashing the 2 or 3 GB of old RAM I would still have room for after adding the new RAM.

    Read the article

  • Is 10% too much for autogrow on a 4 GB sql server DB?

    - by ntsue
    I am getting the following error: 2011-03-07 21:59:35.73 spid64 Autogrow of file 'MYDB_DATA' in database 'MYDB' was cancelled by user or timed out after 16078 milliseconds. Use ALTER DATABASE to set a smaller FILEGROWTH value for this file or to explicitly set a new file size. I did some research, and I found that for large databases you should set autogrow to a fixed size (MB), and not to a percentage. I feel like this database is not large and I may not be addressing the correct issue by changing this value. Does anyone have any opinions? Thank you! EDIT: I should have specified SQL Server 2008 RC2 running on Windows Server 2008

    Read the article

< Previous Page | 414 415 416 417 418 419 420 421 422 423 424 425  | Next Page >