Search Results

Search found 2693 results on 108 pages for 'keeping up'.

Page 42/108 | < Previous Page | 38 39 40 41 42 43 44 45 46 47 48 49  | Next Page >

  • Grouping extensions in Asterisk

    - by Matt H
    We have an asterisk based phone system with multiple clients connected to it. At the moment, all the extension numbers fall in the range 100 - 9999. But, now we have an issue where a particular customer wants to come onto our service and insists on keeping their existing extension list. Is there a way of having extensions 700 for one customer, and also the same number range to another customer. But yet having them belong to a different group? I see there is a concept of groups in asterisk but none of them seem to provide a solution to this.

    Read the article

  • Symantec Backup exec 2012 for SBS - moving data to offsite drives

    - by Will Lennard
    Can anyone advise how best to run the following strategy, or even if it is possible/sensible strategy? I am looking to backup an SBS server using BE2012 where the data is backed up on a weekly full backup, with subsequent daily incrementals till the end of the week, a fresh full backup will then be taken the next weekend, and the cycle repeats for 4 weeks, thus 4 full backups with incrementals. This months worth of backups will then be moved off to an offsite store via USB drive for safe keeping. Is this viable, sensible, or am i just a bit crazy? All ideas are welcome.

    Read the article

  • C drive should only contain OS. Myth or fact?

    - by Fasih Khatib
    So, I have a 500GB HDD @7200RPM. It is split as: C: 97GB D: 179GB E: 188GB My belief is to keep OS ONLY in C:\ and any adamant programs that won't go anywhere apart from C:\ [because this speeds up the PC during startup process] and install programs in D:\ so that in case I have to reinstall the OS, I will have the programs readily available after reinstall. But I have begun to think this approach is flawed because if C:\ is formatted, I will lose registry values and stuff that goes in %appdata% and so it is no use keeping programs in D:/ drive because they will be useless after all. Should I go ahead and install ALL of my programs in C:\ and then use D:\ and E:\ for storing my data like photos, text files, java files n all? How will this impact the performance of the HDD? I only have 3 programs in D:\Program Files so it will be easy to reinstall them :)

    Read the article

  • My NTFS Partition keeps becoming "unusable" on Ubuntu, Any Ideas?

    - by gopherman
    I just purchased a new 2TB Drive External Seagate, My main system uses both Windows and Ubuntu So I am pretty much stuck with keeping my drive as NTFS. I have done this without any problems before but since I got this new drive I have been having issues. When I first load up Ubuntu the drive mounts and runs fine, after an unspecified amount of time i start getting Input/Output errors when accessing the drive. When I goto the Disk Utility I get a message stating the drive is "Unknown or Unused", If I disconnect and reconnect the drive or reboot everything is fine again. There's no errors coming up with S.M.A.R.T and it seems to work fine while under windows. Any thoughts?

    Read the article

  • How to combine RewriteRule of index.php and queries rewrite and avoid Server Error 404?

    - by Binyamin
    Both RewriteRule's works fine, except when used together. 1.Remove all queries except query ?callback=.*: # /api?callback=foo has no rewrite # /whatever?whatever=foo has 301 redirect /whatever RewriteCond %{THE_REQUEST} ^[A-Z]{3,9}\ /([^?#\ ]*)\?[^\ ]*\ HTTP/ [NC] RewriteCond %{REQUEST_URI}?%{QUERY_STRING} !/api(/.*)?\?callback=.* RewriteRule .*$ %{REQUEST_URI}? [R=301,L] 2.Rewrite index.php queries api and url=$1: # /api returns data index.php?api&url= # /api/whatever returns data index.php?api&url=whatever RewriteRule ^api(?:/([^/]*))?$ index.php?api&url=$1 [QSA,L] RewriteRule ^([^.]*)$ index.php?url=$1 [QSA,L] Any valid combination to this RewriteRule's on keeping its functionality? This combination will return Server Error 404 to /api/?callback=foo: # Remove all queries except query "callback" RewriteCond %{THE_REQUEST} ^[A-Z]{3,9}\ /([^?#\ ]*)\?[^\ ]*\ HTTP/ [NC] RewriteCond %{REQUEST_URI}?%{QUERY_STRING} !/api(/.*)?\?callback=.* RewriteRule .*$ %{REQUEST_URI}? [R=301,L] # Rewrite index.php queries RewriteCond %{REQUEST_URI}?%{QUERY_STRING} !/api(/.*)?\?callback=.* # Server Error 404 on /api/?callback=foo and /api/whatever?callback=foo RewriteRule ^api(?:/([^/]*))?$ index.php?api&url=$1 [QSA,L] RewriteCond %{REQUEST_URI}?%{QUERY_STRING} !/api(/.*)?\?callback=.* RewriteRule ^([^.]*)$ index.php?url=$1 [QSA,L]

    Read the article

  • My NTFS Partition keeps becoming "unusable" on Ubuntu, Any Ideas?

    - by gopherman
    I just purchased a new 2TB Drive External Seagate, My main system uses both Windows and Ubuntu So I am pretty much stuck with keeping my drive as NTFS. I have done this without any problems before but since I got this new drive I have been having issues. When I first load up Ubuntu the drive mounts and runs fine, after an unspecified amount of time i start getting Input/Output errors when accessing the drive. When I goto the Disk Utility I get a message stating the drive is "Unknown or Unused", If I disconnect and reconnect the drive or reboot everything is fine again. There's no errors coming up with S.M.A.R.T and it seems to work fine while under windows. Any thoughts?

    Read the article

  • How to secure Apache for shared hosting environment? (chrooting, avoid symlinking...)

    - by Alessio Periloso
    I'm having problems dealing with Apache configuration: the problem is that I want to limit each user to his own docroot (so, a chroot() would be what I'm looking for), but: Mod_chroot works only globally and not for each virtualhost: i have the users in a path like the following one /home/vhosts/xxxxx/domains/domain.tld/public_html (xxxxx is the user), and can't solve the problem chrooting /home/vhosts, because the users would still be allowed to see each other. Using apache-mod-itk would slow down the websites too much, and I'm not sure if it would solve anything Without using any of the previous two, I think the only thing left is avoiding symlinking, not allowing the users to link to something that doesn't belong to them. So, I think I'm going to follow the third point but... how to efficiently avoid symlinking while still keeping mod_rewrite working?! The php has already been chrooted with php-fpm, so my only concern is about Apache itself.

    Read the article

  • Connecting to Aerohive AP's from Laptops running Win. 7 using authentication from a Windows 2008 domain server

    - by user264116
    I have deployed a wireless network using Aerohive access points. 2 of them are set up as radius servers. I want my users to be able to use the same user name and password they use when they log onto our domain. They are able to do this from android devices or computers running Windows 8. It will not work on Windows 7 machines. How do I remedy this situation, keeping in mind that the machines are personal machines not company owned and I will have no way to change their hardware or software.

    Read the article

  • Automatically detecting temperature sensors on startup (Ubuntu 10.10)

    - by dpitch40
    I am very close to achieving my goal of setting up a CPU temperature graph that is displayed in the top panel of my desktop. I have the applet and have gotten it to graph temperatures, which appear to be being sensed correctly. However, my machine doesn't find its temperature sensors by default; I have to run sudo modprobe coretemp for the sensors command to work, then log off and back in before the graph applet starts displaying my temperatures. I am wondering if I can somehow tell the kernel to load the coretemp module on startup so I don't have to keep doing these extra steps. I have tried putting this command in my startup applications, but I think its need for root permission is keeping this from working. Is there a way to set up startup applications with root permission, or some other way to ensure that this module is loaded at startup? If anyone is curious, I'm running 64-bit Ubuntu 10.10 on a Lenovo G770 laptop with a Core i5 processor and the 2.6.35 kernel.

    Read the article

  • Mac OS X desktop background gets reset to "Andromeda Galaxy.jpg" when users log out

    - by smokris
    I'm running Mac OS 10.6 server, and have a set of 10.6 and 10.7 workstations connected to it for authentication and MCX. Users have network profiles (home folder stored on server, via AFP). When users are logged in, they can change their desktop background. So far so good. However, the next time they log in, their desktop background has been reset to "Andromeda Galaxy.jpg". Though MCX is enabled and used to control other settings, MCX is disabled for the Desktop. What is keeping the users' desktop background from being preserved? How can I fix it?

    Read the article

  • Reset All My Passwords

    - by E-rich
    I know there are some great apps out there for keeping track of passwords (see how-do-you-keep-track-of-all-your-passwords if you aren't familiar). I'm interested in additional features for managing my passwords, particularly to reset all my passwords. I can imagine that this feature would be pretty complicated to implement, considering: websites have different methods for resetting a password client applications most likely don't have a means to change a password except for through a graphical user interface (as opposed to providing a command line command) So, I'm curious if there is a solution out there and I just haven't found it. If not, perhaps I will have to start another side project.

    Read the article

  • I'll be setting up a dedicated web server at work soon, my first non hobby server - What should I know?

    - by Rogue Coder
    I've been running my own dedicated server running CentOS and a LAMP stack for 2-3 years now, but it's only been hosting my own websites which aren't super important. However, I will soon be setting up a Linux Webserver and Linux Database Server at work, and I'm wondering what are some important things I should be doing. It's an internal server only, so only people in the company can access it. Should I get a slave server for both of my servers for backups? If I do this, how many backups should I be keeping and how often should those backups be done? Right now on my current server I run a cron job nightly to backup my MySQL databases (Usually 40mb files once compressed), and bi-weekly cron jobs to backup my web root. I just store these files on my local computer via FTP. Also, for an internal server like this, should I look at using LightHTTPD or NginX to increase performance, or will Apache be fine?

    Read the article

  • Cannot schedule task for domain user account

    - by Filburt
    On our Win 2008 machines I can't schedule tasks for domain users because the domain name does not resolve to network name but the AD dc name. The "network name" looks like ABCDEFGE-HIJKLM and the "dc" / "name" would look like ABCDEFGE-HIJKLMN. When selecting the domain user account the account qualifier will look like ABCDEFGE-HIJKLMN\task.user. This results in an "invalid account" error. When however keeping the currently logged in user it will display ABCDEFGE-HIJKLM\current.user. Does this behaviour result from the presumable "illegal" domain name? Is there a workaround for this?

    Read the article

  • Hard drive placement

    - by zm15
    I'm a video editor working with large HD files. I am building a new computer and need some help. I will be running 2 hard drives. One with the operating system and all the programs. And one with all the project files I will be working from. I am keeping these seperate. I will be purchasing a 10k rpm hard drive. So i will have a 10k rpm drive, and a 7200rpm drive. Should I put the OS on the faster drive, or put my working files on the faster drive?

    Read the article

  • Visual Studio Development on Virtual Box, Boot Camp, or VMWare Fusion

    - by Eli
    I currently have a Mac, 2ghz and 2 gigs of ram, running OS X Leopard and Virtual Box with a Windows 7 Pro 32bit virtual machine. Performance on the virtual machine is fine for minor tasks but is very clunky while trying to multi-task or develop in Visual Studio 2008. What would be my best option for being able to use Visual Studio, keeping cost and time in mind? 1) Upgrade ram to 4 gigs ($100). Will this really improve my performance enough to use Visual Studio in a Windows 7 vm? Or am I just wasting time/money? 2) Reinstall/restore Windows 7 disk image as a Boot Camp partition. I assume this should improve my performance, yes? 3) Purchase VMWare fusion instead of VirtualBox. Does Fusion require less resources to run? I am open to any suggestions. Thanks in advance

    Read the article

  • Data recovery after user profile corruption on Windows XP

    - by m68z8mi
    I'm away from home for college, and my computer back home has been having some issues. My dad took it to a computer store, and apparently the user profiles somehow got corrupted, so they're locked out of the computer. This is a Windows XP box, but I changed the default administrator account password, so that backdoor isn't a possibility. Now, that computer's HDD has a whole bunch of data on it which my dad would hate to lose, so I suggested that they take the HDD out, plug it into some other computer, and just copy all the data off that way (keeping in mind that the data itself wasn't encrypted). However, the computer store people said that wouldn't be possible unless they had the administrator account password (which I can't remember for the life of me), and that they'd either have to reformat and reinstall Windows, or else use some complicated sounding recovery process costing a decent amount of money. That sounds like complete BS to me, but I'm not 100% sure about it, so I thought I'd get some more opinions. Could someone more knowledgeable about this stuff suggest a good course of action to take?

    Read the article

  • How to use ssh-agent (and graphic passphrase dialog) in Kubuntu 10.10?

    - by halo
    I have recently switched from Ubuntu to kubuntu, both version 10.10. Unfortunately in KDE the ssh passphrase dialog doesn't work out of the box. Everytime my ssh private key is used I need to enter my password. This is neither secure nor comfortable. I have done comprehensive research on the net but only found out dated documentation. Several things I tried didn't work out. Current status: SSH setup working with direct passphrase input ssh-agent running in X session $SSH_AUTH_SOCK set to ssh-agent's socket in X session How to enable ssh-agent for keeping the passphrase in memory for ~15mins and as a bonus always use a graphical dialog for its input? I use asymmetric SSH key pairs for pushing/pulling in Git VCS logging remotely into different server systems

    Read the article

  • Repairing a TFS 2005 install or starting anew.

    - by Johan Buret
    Following : Installing Team Foundation Server on a shared database instance We did use a shared database instance setup for TFS 2005, and that was not a good idea, because of the Reporting Service dependency. The reporting instance on the server gives error code 404. What works now Basic source code control. We're able to check in and out source code. What doesn't work : Everything else, including : Opening and creating new team projects. Build automation. Internal bug tracking. Goal setup Having a fully working TFS install, and keeping the history. 1) A full install of TFS 2005 on the same server, but within its own database and reporting instance. 2) Using another server might be an option, but it's really not prefered Downtime should be minimum, my colleagues needs to be able to work on the source Readings I've read the MSDN page about moving/restoring TFS 2005, but I'm still unsure about what to do. Thanks in advance for help

    Read the article

  • Create a new domain with the same name of a trusted domain

    - by Russ
    I have a domain blah.com that was aquired a while back by my company foo.com. I set up a two way trust between the two domains, but now I want to move their servers into our forest, while keeping the domain name of blah.com. Is this possible? What things do I need to consider when doing this? I know I can't move the domain from its forest into our forest. blah.com is a 2003 native domain/forest foo.com is a 2008R2 domain/2003 forest.

    Read the article

  • Re-encoding a video file and increasing the size -- does this improve the quality?

    - by Josh
    If I have a video file at 320x240 resolution which I want to re-encode (because I don't like the encoding it's in now) and I also want to play it at double size (640x480), will I get higher quality if I scale it up to 640x480 when I convert it to a new format, verses keeping it at 320x240 in the new format and playing it at double size? This probably depends on the program used to convert, and if so, please let me know any program which might increase the quality. Here's my thinking. If I play a 320x240 file at double size, the system has to scale up each frame in real time, whereas if I scale up while recompressing the system may be able to use a more intensive algorythm like Bicubic interpolation . However I am not sure if this is true or not.

    Read the article

  • How to keep program always in memory (no swapping)

    - by Yossarian
    I'm using KeePass (on Windows 7) for storing passwords. The application is running on my laptop almost 100% of the time, but I'm using it rarely. The problem is, that activating the window after long time takes huge amount of time, I'd say that OS realized the application is no longer used and swapped it. I don't want OS to swap my KeePass (because of the loading lag, and also because of security - I don't like the idea of password keeping program's memory to be stored anywhere on HDD). Is there any possibility how to tell OS not to swap some program? Or, is this achievable by some programming?

    Read the article

  • How to use ssh-agent (and graphic passphrase dialog) in Kubuntu 10.10?

    - by halo
    I have recently switched from Ubuntu to kubuntu, both version 10.10. Unfortunately in KDE the ssh passphrase dialog doesn't work out of the box. Everytime my ssh private key is used I need to enter my password. This is neither secure nor comfortable. I have done comprehensive research on the net but only found out dated documentation. Several things I tried didn't work out. Current status: SSH setup working with direct passphrase input ssh-agent running in X session $SSH_AUTH_SOCK set to ssh-agent's socket in X session How to enable ssh-agent for keeping the passphrase in memory for ~15mins and as a bonus always use a graphical dialog for its input? I use asymmetric SSH key pairs for pushing/pulling in Git VCS logging remotely into different server systems

    Read the article

  • Creating encrypted database for work

    - by Baldur
    My boss posed this problem to me: Encrypted: We need an encrypted database for miscellanious passwords we use at work that are currently only in people's head. Easily accessable: Someone needs to be able to quickly access specific passwords, possibly at hectic moments. This requires any sort of public key management (keeping it on a USB key in a sealed envelope?) to be relatively easy. Access control: The system should have groups of passwords where only specific people have access to specific groups. Recoverability: We need to make sure passwords from one group aren't lost even if the only users with direct access quit or pass away—hence we need some way where (for example) any two members of senior management may override the system (see the treshold link below) and retrieve all the passwords with their key. The first thing that jumped into my mind was some form of threshold and asymmetric cryptography but I don't want to reinvent the wheel, are there any solutions for this? Any software should preferrably be free and open-source.

    Read the article

  • rsnapshot stats

    - by Obscur Moirage
    I'd like to retrieve the following stats from rsnapshot files synced added files modded files deleted files Is there a feature to retrieve these in rsnapshot, or is there another product that's able to do it? EDIT: As requested, I'll try to show that I'm not just asking what I want to do without any research. I wasn't able to locate any rsnapshot feature doing this. Maybe I'm searching in a wrong direction. So, I've built a not very pretty script, called each time before rsnapshot is ran. This Perl script stores each file MD5, in order to compare backup files structures between rsnapshot updates. I'm pretty sure it's worthless to show this code here. I think that keeping an eye on what change on a server, for example, is a useful feature. So, I'm asking. @pauska Most of the time, I'm trying to search for an answer myself, which is not the case here. Thanks

    Read the article

  • Linux router with diffent gateways for incomming and outgoing connections

    - by nkout
    I have the following topology: LAN Users:192.168.1.2 - 254 (192.168.1.0/24) gateway1: 192.168.2.2/24 used for all outgoing connections of LAN users (default gateway) gateway2: 192.168.3.2/24 used for incoming services (destination NAT, ports 80,443 are forwarded to 192.168.2.1) linux router-server R eth0 192.168.1.1/24: LAN eth1 192.168.2.1/24: WWAN1 eth2 192.168.3.1/24: WWAN2 I want to: route all outgoing traffic coming from LAN and R via 192.168.2.2 route the responses to incoming connections via 192.168.3.2 My config: ifconfig eth0 up 192.168.1.1 netmask 255.255.255.0 ifconfig eth1 up 192.168.2.1 netmask 255.255.255.0 ifconfig eth2 up 192.168.3.1 netmask 255.255.255.0 echo 0 >/proc/sys/net/ipv4/ip_forward route add default gw 192.168.2.2 iptables -t nat -A POSTROUTING -d !192.168.0.0/16 -j MASQUERADE I want to add iptables rule to mark incoming traffic from WWAN2 and send back the responses to WWAN2, while keeping default gateway on WWAN1

    Read the article

< Previous Page | 38 39 40 41 42 43 44 45 46 47 48 49  | Next Page >