Search Results

Search found 16628 results on 666 pages for 'setup kit'.

Page 393/666 | < Previous Page | 389 390 391 392 393 394 395 396 397 398 399 400  | Next Page >

  • Can I run my MacBook in clamshell mode without being connected to power?

    - by kch
    Hi, At home, I run my MacBook in clamshell mode (closed lid, external display). This works fine when you're connected to the power adapter, but it doesn't work when running on battery. That's how it's supposed to be and Apple has some kb entry on the issue. But it's also lame. You can prevent the machine from sleeping when closed by running InsomniaX, but then it'll assume the builtin display is still active, so you end up with a two-display setup when you really only want the external. This is obviously less than ideal. So, is there any work around, hack, utility, black magic that I can use to make it run in clamshell mode while strictly on battery power? Also, bonus points for a solution that makes the AC status not affect the machine state at all. (Like, you know, it does normally, when not in clamshell.)

    Read the article

  • Software to replicate one computers display onto many other displays

    - by Joe Taylor
    We have a classroom setup with one teachers pc at the front. I am looking for some software, preferably open source although this is not a deal breaker, to force all displays in the room to replicate the teachers display. Also if this software could be locked so the students could not exit this software while it was running. Does anyone know of any software that could perform this task? I have googled around for a solution but haven't found anything suitable as yet. It would be running on Windows 7 Flavours of the software I have found are: Lanschool and NetOp. Open source alternatives would be better.

    Read the article

  • SSO solution and centralized user mgmt for about 10-30 Ubuntu machines?

    - by nbr
    Hello, I'm looking for a clean way to centralize user management. The setup: About 10-30 linux machines (Ubuntu 10.04 LTS server) Maybe 10-30 users for now. The requirements (hopes and expectations): A single place for the administrator to manage user accounts, passwords and the list of machines each user has access to. (And probably groups.) Doesn't have to be fancy. Single sign-on for SSH: the user should be able to login from machine A to machine B without re-entering his/her password. A Quick Google searches give me pointers to OpenLDAP and Kerberos, but I'm not sure where to start and what problem will each solution actually solve. Which way to go? I'd love to find a clear that focuses on this subject. (Or: am I asking "a wrong question"?)

    Read the article

  • CentOS - Apache and SSL?

    - by NeoNmaN
    Hei all i have problem for my SSL setup, when i put SSL Engine to on my Apache will not start up, if i remove it, its work fine. i can't finde the Apache log, so i can find out of whats happen here, can sommen maby tell me the problem or link to the defualt log-path for Apache? here is my vhost config file. <VirtualHost *:443> # SSLEngine On SSLProtocol all SSLCertificateFile "/etc/pki/tls/schemecloud.co.uk.csr" SSLCertificateKeyFile "/etc/pki/tls/schemecloud.co.uk.key" SSLCACertificateFile "/etc/pki/tls/rapidssl.crt" ErrorLog "/var/www/error_ssl_log" <Directory /var/www/html> AllowOverride All </Directory> DocumentRoot /var/www/html/ ServerName schemecloud.co.uk </VirtualHost>

    Read the article

  • Evaluate a vendor laptop before deployment to user?

    - by NetWarrior
    I get numerous requests from executives and users for new smaller laptops for travel purposes. Most of my evaluation is based upon whether or not it can run certain applications. Mainly lotus notes, office, and video. Most of the laptops include windows 7 OS, and are fully loaded with ram, a high-end processor and a integrated graphics card. My boss whats me to document the usefulness of the laptop and performance. I'm just a little confused on how to setup a document that can be used by members of the IT department for future evaluations.

    Read the article

  • Subscribe to feed in Thunderbird from the command line?

    - by Coderer
    From reading around the web, it looks like Firefox's "quick view" of an RSS feed sometimes lets you "Subscribe to this feed using" Thunderbird. For whatever reason, that's not an automatically-added option with my setup (FF 3.5.something + Thunderbird 3.0.something on Linux), so I figured I could just "Choose Application...", point at the Thunderbird binary, and be on my way. Not so -- nothing appears to happen. If I run thunderbird from the command line as thunderbird "http://path/to/feed" the app launches as normal. If it's already running, absolutely nothing happens. Is this impossible? Is there some mojo I can pass Firefox to tell it that Thunderbird exists? Should I just suck it up and copy/paste the URLs manually?

    Read the article

  • Ubuntu 12.04 - syslog showing "SGI XFS with ACLs, security attributes, realtime, large block/inode numbers, no debug enabled"

    - by Tom G
    I have been seeing these random logs in syslog on our production system. There is no XFS setup. Fstab only shows local partitions, only EXT3 . There is nothing in crontabs either. The only file system related package I have installed is 'nfs-kernel-server' Kernel version is 3.2.0-31-generic . kernel: [601730.795990] SGI XFS with ACLs, security attributes, realtime, large block/inode numbers, no debug enabled kernel: [601730.798710] SGI XFS Quota Management subsystem kernel: [601730.828493] JFS: nTxBlock = 8192, nTxLock = 65536 kernel: [601730.897024] NTFS driver 2.1.30 [Flags: R/O MODULE]. kernel: [601730.964412] QNX4 filesystem 0.2.3 registered. kernel: [601731.035679] Btrfs loaded os-prober: debug: running /usr/lib/os-probes/mounted/10freedos on mounted /dev/vda1 10freedos: debug: /dev/vda1 is not a FAT partition: exiting os-prober: debug: running /usr/lib/os-probes/mounted/10qnx on mounted /dev/vda1 10qnx: debug: /dev/vda1 is not a QNX4 partition: exiting os-prober: debug: running /usr/lib/os-probes/mounted/20macosx on mounted /dev/vda1 macosx-prober: debug: /dev/vda1 is not an HFS+ partition: exiting os-prober: debug: running /usr/lib/os-probes/mounted/20microsoft on mounted /dev/vda1 20microsoft: debug: /dev/vda1 is not a MS partition: exiting os-prober: debug: running /usr/lib/os-probes/mounted/30utility on mounted /dev/vda1 30utility: debug: /dev/vda1 is not a FAT partition: exiting os-prober: debug: running /usr/lib/os-probes/mounted/40lsb on mounted /dev/vda1 debug: running /usr/lib/os-probes/mounted/70hurd on mounted /dev/vda1 debug: running /usr/lib/os-probes/mounted/80minix on mounted /dev/vda1 debug: running /usr/lib/os-probes/mounted/83haiku on mounted /dev/vda1 83haiku: debug: /dev/vda1 is not a BeFS partition: exiting os-prober: debug: running /usr/lib/os-probes/mounted/90bsd-distro on mounted /dev/vda1 83haikuos-prober: debug: running /usr/lib/os-probes/mounted/90linux-distro on mounted /dev/vda1 os-prober: debug: running /usr/lib/os-probes/mounted/90solaris on mounted /dev/vda1 os-prober: debug: /dev/vda2: is active swap Why would this randomly show up? This also spawns multiple "jfsCommit" processes.

    Read the article

  • Fortigate VPN Routing issue

    - by user1571299
    I have 200B Fortigate unit with 2 internet WAN connections. I also have a remote site which I'm connected to via IPSEC VPN through WAN1. This site has only one GW IP address. I'd also like to setup a VPN ontop of WAN2 with that specific site as it's destination. The default route for my end is WAN1. My problem is I cant figure out how to have both tunnels up at the same time. What's the best practice for achieving this? Thanks

    Read the article

  • I Can't run the Netbeans but I installed successfully

    - by David
    I'm new to Ubuntu as well as Netbeans. I installed Netbeans, and I've made sure to install all the JDKs and JREs I could find. It installed without errors. I also saw this question and made sure I followed all the instructions there as well. I never got any error messages of any kind. So far as I know, it installed okay. However, when I try to run Netbeans, I get the message in the bottom of the Netbeans IDE like this: ant -f /root/NetBeansProjects/samp1 -Djsp.includes=/root/NetBeansProjects/samp1/build/web/one.jsp -DforceRedeploy=false -Dclient.urlPart=/one.jsp -Ddirectory.deployment.supported=true -Djavac.jsp.includes=org/apache/jsp/one_jsp.java -Dnb.wait.for.caches=true run /root/NetBeansProjects/samp1/nbproject/build-impl.xml:774: The libs.CopyLibs.classpath property is not set up. This property must point to org-netbeans-modules-java-j2seproject-copylibstask.jar file which is part of NetBeans IDE installation and is usually located at <netbeans_installation>/java<version>/ant/extra folder. Either open the project in the IDE and make sure CopyLibs library exists or setup the property manually. For example like this: ant -Dlibs.CopyLibs.classpath=a/path/to/org-netbeans-modules-java-j2seproject-copylibstask.jar BUILD FAILED (total time: 0 seconds)

    Read the article

  • UFW blocking random packets on 443

    - by s2jcpete
    All, I have UFW setup to allow traffic on port 443. It works as expected, though I have a large amount of UFW Block log entries. To Action From -- ------ ---- 80 ALLOW Anywhere 443 ALLOW Anywhere 22222 ALLOW Anywhere 80 ALLOW Anywhere (v6) 443 ALLOW Anywhere (v6) 22222 ALLOW Anywhere (v6) However in my syslog file I see this: [UFW BLOCK] IN=eth0 OUT= MAC=XXX SRC=<foreignip> DST=<serverip> LEN=40 TOS=0x00 PREC=0x00 TTL=116 ID=22025 DF PROTO=TCP SPT=49622 DPT=443 WINDOW=0 RES=0x00 ACK RST URGP=0 About 30 or so seconds later pound (which I'm using for SSL decryption and port redirection) throws a connection timed out messsage. I'm assuming this is because UFW is blocking the packet. I'm at a loss as to an explination. Could the packet be malformed or something, is this normal? Edit - I have since changed the /etc/defaults/ufw and set ipv6=no, so the v6 rules are no longer in the mix. The server is still showing the block / connection timed out behavior though. The new ufw status output is: Status: active Logging: on (low) Default: deny (incoming), allow (outgoing) New profiles: skip To Action From -- ------ ---- 80 ALLOW IN Anywhere 443 ALLOW IN Anywhere 22222 ALLOW IN Anywhere

    Read the article

  • Developing and implementing a testing plan for a software app deployed on a web server

    - by Abhzoo
    A company in the USA is building a new Web App that will be offered SaaS to customers and the development is being done by a software development team located in a different country(India). They are about to take delivery of a first demo to provide live feedback to the team in India. The overseas team requires a cloud server (Windows + SQL Standard, 8GB Ram, 8 vCPUs, 40GB SSD system disk, 80GB SSD data disk, 1600Mb/s network bandwidth) to serve as a tester server. When the tester is setup the team will install the app on the test server to get live feedback. Q:Explain in detail how you will develop and implement a testing plan for the software App. Be sure to explain the specifics. PLEASE HELP, NEED ANSWER ASAP

    Read the article

  • Apache expands REQUEST_FILENAME with filetype

    - by Lauer
    I have a .htaccess fil on a Apache22 server. RewriteEngine On RewriteBase / RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteRule ^(.+)$ /index.php?p=$1 [L] The idea is when the request is /testpage it should be rewritten to /index.php?p=testpage. The problem is that apache does a write before my file, so the actually request is testpage.xml which actually does exists. Therefore, the first RewriteCond is newer true, and my rewrite rule is never applied on the /testpage request. There server should be a standard setup on a FreeBSD server. I know this work on other servers.

    Read the article

  • Nginx proxy to Apache - resolve HTTP ORIGIN

    - by Fratyr
    I have a server setup with nginx serving static content and proxy all PHP/dynamic requests to apache on 127.0.0.1 I'm building an API for my databases, and I need to allow clients by their origin (domain name), rather than just IP. Based on CORS rules. So when I send an HTTP header header("Access-Control-Allow-Origin: www.client-requesting.myapi.com"); from my API server, I have to tell it which origin I allow, otherwise client side requests won't work to my API due to same-origin policy. The question is how can I know which domain name (if any) called my API? What should be the nginx and apache configuration to pass the origin parameter? I tried to google, and all I found is some possible solution with mod_rpaf, but I wanted to be sure. Thanks!

    Read the article

  • Getting VPN to work from Virtual PC

    - by strobaek
    Setup: Host is running Windows7. Virtual PC is a Windows Server 2008 running under VMWare workstation 6.5. From the Host I have a VPN to e.g. TFS and other resources. From the VPC I need to connect to e.g. a SQL server via the VPN. My problem is, that I cannot get a connection from the VPC. If I'm sitting on the corporate network, all is working fine (but then I don't have the VPN). From home - where the VPN is required - it does not work. I have two network adapters defined/configued. One as BRIDGED and one as Host Only. IF I change the one being BRIDTED to NATS I have no connectivity at all from the VPC. I have no problems connecting from my host to the VPC. Thanks.

    Read the article

  • Unsure about TRIM enabled on my SSD

    - by user84750
    I have a SSD OCZ Vertex4 installed on my laptop. I'm running Ubuntu 12.04 LTS. I have enable TRIM by adding "discard" to my fstab file. (also added option noatime). I rebooted my Ubuntu and followed These instructions here to test TRIM. The end results of my tempfile was all ffff's, when it should have read all zero's, which is telling me TRIM is not really working or enabled correctly. Did I miss something? Also, will it be a problem if only my /home directory is encrypted. AND if you ask why I have swap on my SSD, it's because I let Ubuntu set up my partition. When I have my SSD, I just wanted to install Ubuntu as fast as possible. =) I've done testing to see at which point it will start to use swap and it took a lot of applications open to finally use swap. I currently have 4 GB of memory. I might shrink this to like 512 MB or 1 GB the most. Here's some info about my file system setup. sudo hdparm -I /dev/sda1 | grep "TRIM supported" Data Set Management TRIM supported (limit 16 blocks) sudo fdisk -l /dev/sda Device Boot Start End Blocks Id System /dev/sda1 * 2048 242016255 121007104 83 Linux /dev/sda2 242018302 250068991 4025345 5 Extended /dev/sda5 242018304 250068991 4025344 82 Linux swap / Solaris ls /dev/mapper control cryptswap1

    Read the article

  • Virtualizing Firewalls/UTM

    - by TheD
    So Sophos UTM (Astaro Gateway UTM) can be virtualized in Hyper-V, VMWare, vSphere etc. Now I'm new to virtualization, so am wondering how the networking would be configured in this situation. Imagine the following (a very simple setup). A Zyxel router acting as the default gateway, standard ADSL line. This is connected directly via Ethernet to my VHost, Hyper-V in my case. Now I assume I need two virtual network adapters, one which is the "External" interface to the firewall, and the other being the "Internal" interface. But where does my switch go, since I have 5 machines (or X number of machines). Where is the switch connected, to the internal interface on the VHost? What default gateway do my clients use etc? How would the routing work if these two interfaces weren't bridged since they are virtual networks anyway. I'm a little confused! Any advice is much appreciated! Cheers

    Read the article

  • How to connect two subnet with Windows Server?

    - by 9dan
    Hello, I have some work to do in a small office. This office has two separate Internet connections. One is a kind of DSL line and connected through the Internet sharing router (DHCP). One is a leased line with public IP adresses (No DHCP). This leased line provide only 30 IP addresses and the gateway is provided by the ISP. Some workers use private IP provided by the router, some workers use manually setup public IP. There is a Windows 2008 R2 server with two NICs, connected to each lines. I want to connect/merge(?)/bridge(?) two subnet with this server so that users from the different subnet can access each other. How can I accomplish this? Something like, add gateway function to this server and public IP users change the gateway to point this server.. It it possible ? Sorry for my rambling. Thanks in advance.

    Read the article

  • How to troubleshoot slow powerconnect 62xx management interface

    - by Hannes
    Our Dell Powerconnect 62xx switches have a very high packetloss on the management interface. I presume this is caused by a new appliance which uses multicast for communication but I am not sure. Our network setup is following: servers a - Dell PC6248 | servers b - Dell PC6248 |- juniper core router servers c - Dell PC6248 | What we see is that the multicast traffic arrives at all servers (but only the servers b use the multicast) and I fear that this multicast traffic floods the switch management interface. The switches' management interfaces are reachable via vlan101, all other traffic is sent over other vlans. When I tcpdump on one of the 2 servers with a vlan 101 ip address, I only get a few arp requests but almost nothing. When I try to ping between these 2 servers, it works like a charm. I would like to know what a good way is to troubleshoot this problem and maybe help me understand what is going wrong on that subnet.

    Read the article

  • Windows Server 2008 r2 FTP blocking outside connections

    - by nbon
    I have a windows server 2008r2 running IIS 7.5. I am trying to setup a FTP-server in IIS but I'm running into some annoying problems. Setting up the server works fine but when I try to connect from a remote client the connection times out. I have tried to connect to the FTP-server from the localhost and it works flawlessly. I figured that it should be some trouble with the firewall so I went into firewall settings and disabled the Public Profile and my remote connections worked! In my inbound rules there are rules for FTP-connections to allow all profiles etc. I guess they are made automatically when setting up the FTP-server. Anyone got any idea how to allow remote connections without turning off the public firewall?

    Read the article

  • suppress warnings from windows live mail untrusted cerficate ssl

    - by monkeyking
    Hi we have setup our own mailserver using ssl. Each time I start up windows live mail, in comes up with an annoying message that "The server you are connected to is using a security certificate that could not be verified" Is there some way to add this cerficate such that we wont see this message on every program start. When using firefox (our mailserver can also be webbased) or thunderbird, I get the option to allow the certificate without asking again. I'm perfectly aware that we can buy a certificate that will make the message go away, but this we dont want to do. Thanks edit: I have succesfully imported the certificate such that I can access the website without complains using ie, however the problem still persists in windows live mail.

    Read the article

  • Chef cookbooks marked as executed

    - by Gonzalo Alvarez
    I have setup a Chef server in our network that I use to manage several nodes. These nodes have a chef-client installed executing as a daemon every X minutes. The problem is that every time the client runs, it executes the recipes for all the cookbooks, even those previously executed, so it consumes resources and sometimes it even breaks thinks (with services restarts, for example). I know that I can avoid executing a piece of code or a recipe I create as detailed here: Prevent chef recipe from executing previously executed action? but, would this mean that I should modify any cookbook that I download from the opscode repository. In other words, is it possible to make the chef server (or the clients) to mark the cookbooks as 'executed' as soon as they are executed the first time?

    Read the article

  • WebLogic Partner Community Newsletter October 2013

    - by JuergenKress
    Dear WebLogic Partner Community member, Our October newsletter edition focuses on Oracle OpenWorld 2013, highlights, keynotes and all presentations. Thanks to all the partners who made the conference a huge success, if you could not come to San Francisco, you can find all the details in this newsletter. We added additional locations for the free hands-on ADF & ADF Mobile Bootcamps & WebLogic Bootcamps. As a community member you can also get a free voucher to become a WebLogic Server 12c Certified Implementation Specialist or ADF 11g Certified Implementation Specialist (limited to partners from EMEA!) If you can not make it to a Bootcamp, do not miss the virtual developer days for WebLogic and ADF Mobile. If you plan to install WebLogic read first the article “Setup a 12c Fusion Middleware Infrastructure from René van Wijk. If you administrate Middleware make sure you read the documentation and support notes Weblogic Server Patching & Maintenance Information Center. In the ADF section of the newsletter our product management team continues with the ADF Architecture on-demand training. Andrejus released the latest version of the ADF Performance Audit Tool v 2.0. The summer is over, if you look for a Christmas present, for your kids or yourself maybe you want to run Java on LEGO® Mindstorms® EV3. Jürgen Kress To read the newsletter please visit http://tinyurl.com/WebLogicNewsOctober2013 (OPN Account required) To become a member of the WebLogic Partner Community please register at http://www.oracle.com/partners/goto/wls-emea ( OPN account required). If you need support with your account please contact the Oracle Partner Business Center. Blog Twitter LinkedIn Mix Forum Wiki Technorati Tags: WebLogic Community newsletter,newsletter,WebLogic,WebLogic Community,Oracle,OPN,Jürgen Kress

    Read the article

  • TF2010 Build Definition and Access to Path is Denied error?

    - by Daniel DiVita
    I am new to TFS with regards to build definitions. I have a a build folder setup where I have set the permissons so EVERYONE has full control. Here is the exact error I am getting: E:\Builds\PIMSite\PIM.Site\PIM_Site.metaproj: Unable to copy file "C:\Builds\1\PIM System\PIM Site Build\Binaries\HtmlAgilityPack.xml" to "..\PIM.Site\Bin\HtmlAgilityPack.xml". Access to the path '..\PIM.Site\Bin\HtmlAgilityPack.xml' is denied. I have tried everyhitng. I have removed everything from that folder adn can delete it just fine so it is not being used by another process. Any thoughts?

    Read the article

  • Migrate TFS 2010 Application Tier to another server on the same domain

    - by Liam
    I'm in the process of looking into the possibility of moving our TFS 2010 Application Tier to another server from the one it is on at the moment so that we can repurpose the hardware. I've been looking through the Microsoft Documentation over at http://msdn.microsoft.com/en-us/library/ms404869.aspx, but this assumes that everything is stored on one box (Application and Data tiers). In my setup however, our Data tier is separate to the Application tier and will be staying where it is. I think I should be able to do this but for my own peace of mind, would there be any issues or implications if I merely installed the Application Tier on the new hardware and then connected it to the existing data tier?

    Read the article

  • How to copy protected files when an Administrator in Vista (easily)

    - by earlz
    Hello, I have a harddrive I need to backup. In the harddrive is of course things like Documents and Settings which is set to not allow other people to see inside someone's personal folders. I am an administrator though and I can not figure out how to mark these files so that I am permitted to access them and copy them. IWhen I double click on My Documents then it pops up saying You must have permission to access this and gives me an option like ok or cancel. I click ok and then it says you do not have permission to access these files I'm an administrator on the system so I don't understand why Vista is locking me out. How can I setup vista so that it will let me copy every file, even ones I don't have permission to?

    Read the article

< Previous Page | 389 390 391 392 393 394 395 396 397 398 399 400  | Next Page >