Search Results

Search found 21717 results on 869 pages for 'setup versions'.

Page 547/869 | < Previous Page | 543 544 545 546 547 548 549 550 551 552 553 554  | Next Page >

  • How to reference a git repository?

    - by Anonymous
    What should the actual path to a git repository 'file' be? It's common practise when cloning from github to do something like: git clone https://[email protected]/repo.git And I'm used to that. If I init a repo on my local machine using git init, what is the 'git file' for me to reference? I'm trying tot setup Capifony with a Symfony2 project and need to set the repository path. Specifying the folder of the repository isn't working. Is there a .git file for each repository I should be referencing?

    Read the article

  • Sudden complaint that Windows is not genuine - some questions

    - by blade
    Hi, I used to use VMWare Workstation and made several VMs with Win Server 2008, which worked fine. My first tasks were to setup RDP, updates, and activate the OS. I deleted Workstation and thus did not have my VMs available to use for about 2 months. Now I have VMWare Server 2.0.2 as I this is free and I am going to get Hyper-V on a host Win Server in about March, but when I have accessed one of my VMs (haven't tried the others yet), it complains that the copy of Windows is not genuine. I now have a black screen but can access all my apps on the VM. Will the OS eventually shut me out from accessing it? I got my key from the Action Pack but the standard R2 key has been taken down and replaced with the Enteprise R2 key. I have Enterprise R2 but apparently the autorun is corrupt! Thanks

    Read the article

  • PCs with Wired Connection Keep Losing Internet Connectivity

    - by user717452
    I have a U-verse ATT DSL Internet at our church building. Recently, any computer that is plugged into the network via Ethernet keeps losing its internet connectivity. The setup is from the Modem to a Netgear Wireless Router, and from the router to 3 different computers. None of the laptops that use Wifi ever lose internet, just the wired ones, and I end up having to do ipconfig /release and /renew every day to get it back. One computer has Windows 7 and one has Windows XP. Any ideas as to what is going on with our network?

    Read the article

  • Planning office network [closed]

    - by gakhov
    I'm planning to setup my office network from scratch and want to ask professional opinions or tips. My office is connected to Internet with Cable connection (100Mb/s). The devices i would like to connect are VoIP Phone (RJ-11), TV (WiFi/LAN), 3 laptops (WiFi), a few smartphones (WiFi), iPad (WiFi), Kindle (WiFi) and, probably, MediaServer (WiFi/LAN). As you can see, the most load will be on WiFi connections (probably, even if TV supports WiFi it's better to connect it by LAN?). So, i need help to choose the best routers combination (or even one?) to support stable connections for all these devices and minimize the total number of routers/adapters. Any thoughts? Thank you!

    Read the article

  • Does OEM partition stores Windows 7 serial key ?

    - by Mahesha999
    Does OEM partition stores the serial key which is grabbed by the Win 7 setup done from the disk provided with laptop. I realized that when you install clean system from Windows 7 disk you get with laptop, it grabs the serial key or it doesn't ask one. So from where it gets that serial key. Is it the OEM partition or anything else? Or the recovery partition stores the key? What all the things these two partitions do store in them?

    Read the article

  • Downloading a repository for local use

    - by EBV2010
    I'm trying to get Thunderbird working in such a way that it will properly work with Kolab groupware. For that I need it to be in a fixed setup of Thunderbird and add-ons (Lightning, SyncKolab) without automatic updates and I need to present version of Thunderbird to be available for the users. What I hope to achieve is that the repository for Thunderbird as it is now on http://ppa.launchpad.net/mozillateam/thunderbird-stable/ will be available on my local server so I always use that version even if Thunderbird goes to a new stable version. What I hope to achieve is this: - I copy the content of http://ppa.launchpad.net/mozillateam/thunderbird-stable/ to my server - I make it available as a repository on my network I neither know if this is possible or allowed under the license etc.

    Read the article

  • Frame rate on one of two machines running same code seems to be capped at 60 for no reason

    - by dennmat
    ISSUE I recently moved a project from my laptop to my desktop(machine info below). On my laptop the exact same code displays the fps(and ms/f) correctly. On my desktop it does not. What I mean by this is on the laptop it will display 300 fps(for example) where on my desktop it will show only up to 60. If I add 100 objects to the game on the laptop I'll see my frame rate drop accordingly; the same test on the desktop results in no change and the frames stay at 60. It takes a lot(~300) entities before I'll see a frame drop on the desktop, then it will descend. It seems as though its "theoretical" frames would be 400 or 500 but will never actually get to that and only do 60 until there's too much to handle at 60. This 60 frame cap is coming from no where. I'm not doing any frame limiting myself. It seems like something external is limiting my loop iterations on the desktop, but for the last couple days I've been scratching my head trying to figure out how to debug this. SETUPS Desktop: Visual Studio Express 2012 Windows 7 Ultimate 64-bit Laptop: Visual Studio Express 2010 Windows 7 Ultimate 64-bit The libraries(allegro, box2d) are the same versions on both setups. CODE Main Loop: while(!abort) { frameTime = al_get_time(); if (frameTime - lastTime >= 1.0) { lastFps = fps/(frameTime - lastTime); lastTime = frameTime; avgMspf = cumMspf/fps; cumMspf = 0.0; fps = 0; } /** DRAWING/UPDATE CODE **/ fps++; cumMspf += al_get_time() - frameTime; } Note: There is no blocking code in the loop at any point. Where I'm at My understanding of al_get_time() is that it can return different resolutions depending on the system. However the resolution is never worse than seconds, and the double is represented as [seconds].[finer-resolution] and seeing as I'm only checking for a whole second al_get_time() shouldn't be responsible. My project settings and compiler options are the same. And I promise its the same code on both machines. My googling really didn't help me much, and although technically it's not that big of a deal. I'd really like to figure this out or perhaps have it explained, whichever comes first. Even just an idea of how to go about figuring out possible causes, because I'm out of ideas. Any help at all is greatly appreciated.

    Read the article

  • PackageForTheWeb "should not be executed directly"

    - by Sergei Ousynin
    When trying to install package (specifically, chipset drivers for Nexcom PEAK-series computer, but I guess it doesn't really matters) I get "PackageForTheWeb Error" saing "This program is used internally by PackageFromTheWeb. It should not be executed directly." I get this error on both Windows 2000 (where it should be installed) and Windows 7. The original driver CD was lost long time ago, so the package was downloaded from the Nexcom website, found no info about installer prerequisites there. What should I install (or anything else) to install this package? UPD: I found a solution - unpacked the package and executed setup.exe that was inside. But I'm pretty sure it's not an action supposed by vendor, so answers are still welcome.

    Read the article

  • Windows Vista Home memory usage problem

    - by lordg
    Hi, I have a Windows Vista Home laptop from a client that is running on 1GB ram. The laptop is used for super basic things, word, internet, outlook, etc. What makes zero sense is that the RAM is being completely consumed, causing the PC to hang sometimes when it can't take it anymore. However, in task manager, the processes appear to only be consuming maybe 100MB (Private Working Set). The client literally has a simple setup, and is running kaspersky, though that does not seem to be indicating it is the cause of the excessive memory usage. Does anyone have a suggestion on how to resolve the memory issue or how to track down what is actually happening and fix it? G

    Read the article

  • Server App - Users

    - by Benno
    I'm using mac OSX lion server, and I'm trying to setup a user account with access to different services. I had to create the user in Prefs users and groups, because it wouldn't work in the Server app. In the server app, it kept saying "This operation couldn't be completed". I managed to create the user in the system prefs, but now I'm back in the server app and cannot update the services the user has access to. The checkboxes are all there for e.g. Address book, file sharing etc, but I can't deselect any of them... Any ideas on why I can't change a user's access to services?

    Read the article

  • How do I make virtual host DirectoryIndex file appear in the url?

    - by Bob Flemming
    I have setup a virtual host which specifies a default file to load when the URL is called. The problem I have is that I need that default DirectoryIndex file to appear in the URL. So when I go to: www.mysite.co.uk, I want www.mysite.co.uk/app.php to appear in the URL. How can I achieve this using my virtual host configuration within my apache.conf file? Here is my current code: <VirtualHost *:80> ServerName *.mysite.co.uk DocumentRoot "/var/www/html/mysite/web/" DirectoryIndex app.php </VirtualHost>

    Read the article

  • Booting Ubuntu 12.04 from external eSATA disk

    - by Lord of Scripts
    This is my system topology: Disk #1 (SATA Internal) C: D: (Windows 7 Ultimate) Disk #2 (SATA Internal) E: (Windows Backup) Disk #3 (eSATA External) H: I: (Other windows data) /dev/sdc3 Linux Swap /dev/sdc4 Extended partition /dev/sdc5 Linux / So, I originally had there Ubuntu 8.1 from years ago but never got to use it. Now I used the Ubuntu 12.04 Live CD to install on that same location (That live CD takes a century to boot on a 6GB Intel i7 system...). The installation went fine, I selected it to install on /dev/sdc5 but it never asked me for any boot stuff, where I wanted to install Grub or whatever it is that it uses nowaways (I come from the LILO days when it always worked :-) So, yet again I can't access my new Linux installation. I have to wait a century to boot the "Live" CD and it allows me to see my new installation but I can't do anything with it. I tried the approach of this blog post. Copied the linux.bin of /dev/sdc5 into C: and used the BCDEdit steps to declare the new OS. So when I boot I see the Windows Boot menu and select Linux and after than I only get a black screen with a blinking cursor on the upper left. I can boot into Windows though. So, perhaps it didn't install the boot code on /dev/sdc5? I used this setup years ago booting from Windows with a BIN file: dd if=/dev/sdc5 of=/mnt/share/C/linux.bin bs=512 count=1 I am very reluctant to run GRUB because years ago I did and it wiped out my Windows boot sector and took quite some effort to recover it and be able to boot Windows again. I have been trying to install GRUB on a blank USB stick but I can't find anything clear enough. My system does NOT have a floppy. So can someone give me some ideas about how to get control of my Ubuntu 12.04 installation?

    Read the article

  • Web log files analyzer

    - by Peter Štibraný
    I already use Google Analytics on my page, but I'd like to get additional info from log files. I've looked at various packages during last days, but nothing impressed me so far. Some requirements: must work on log file level (I use apache combined logs, but can configure apache to produce other types of logs) can generate static reports (windows/linux) or use GUI (windows only) should be easy to add custom user agents, and rerun analysis if it can recognize installation of eclipse plugins from log, that would be big plus understands google serp position referer should not require two days to setup (awstats, I am looking at you) should be still under active developement (i.e. analog isn't good answer) preferrably free, or at not very expensive :-) Any good analyzers programs out there?

    Read the article

  • /dev/shm (shared memory) on linux

    - by Kirzilla
    Hello, Let's imagine that we have 8Gb of RAM on server. I'm mounting /dev/shm with 4Gb on board. mount -o remount,size=4G /dev/shm Will this memory be strictly reserved for shared memory or if /dev/shm is empty this memory could be used by regular applications (web server, php etc.)? PS:Sorry for my English. I'm asking it because I've just checked df -h and found tmpfs 6.0G 0 6.0G 0% /dev/shm on 8Gb RAM sever. I don't know who made this setup, but it seems to me awful. Thank you!

    Read the article

  • How to define implementation details?

    - by woni
    In our project, an assembly combines logic for the IoC-Container, the project internals and the communication layer. The current version evolved to have only internal classes in addin assemblies. My main problem with this approach is, that the entry point is only available over the IoC-Container. It is not possible to use anything else than reflection to initialize the assembly. Everything behind the IoC-Interface is defined as implementation detail and therefore not intended for usages outside. It is well known that you should not test implementation detail (such as private and internal methods), because they should be tested through the public interface. It is also well known, that your tests should not use the IoC-Container to setup the SUTs, because that would result in too much dependencies. So we are using the InternalsVisibleTo-Attribute to make internals visible to our test assemblies and test the so called implementation details. I recognized that one problem could be the mixup between different concerns in that assembly, changing this would make this discussion useless, because classes have to be defined public. Ignoring my concerns with this, isn't the need to test a class enough reason to make it public, the usages of InternalsVisibleTo seems unintended, and a little bit "hacky". The approach to test only against the publicly available IoC-Container is too costly and would result in integration style tests. The pros of using internals are, that the usages are well known and do not have to be implemented like a public method would have to be (documentation, completeness, versioning,...). Is there a solution, to not test against internals, but keep their advantages over public classes, or do we have to redefine what an implementation detail is.

    Read the article

  • Wordpress 3 multi-site install

    - by mike
    Hello, Trying to figure out if this is possible... My company has a cms product that was written in Java and we decided to use Wordpress to run blogs for our clients. Obviously, Wordpress does not run on tomcat(at least not by default) so we installed Pound(http://www.apsis.ch/pound/) on our server and have setup any Apache and Tomcat on different ports. When "/blog/" is requested, the request is directed to Apache. This works fine but we would like to use Wordpress multi site so that we can manage all the blogs from a single interface. We would also like the url for every site to be "/blog/" example: http://www.site1.com/blog/ http://www.site2.com/blog/ I'm thinking it would have to be done with apache??? Is it even possible? Thanks!

    Read the article

  • Windows 8.1 VPN Connection Crash

    - by Anthony Russell
    I am attempting to connect to a VPN that I have used for quite some time. I followed these instructions to setup my VPN on both my desktop and my surface pro Both my desktop and surface pro have all of the current updates for Windows 8.1. However, my surface pro connects to the VPN without issue and my desktop crashes when I attempt to do ANYTHING VPN related. If I Attempt to connect to a VPN it freezes and crashes without error as soon as I hit connect Attempt to remove a VPN it freezes and crashes without error as soon as I hit remove Attempt to connect to a newly created VPN it freezes and crashes without error as soon as I hit connect I have tried disabling all Windows Firewall proticols. I have tried disabling the Windows Defender antimalware shenanigans I have no other antimalware, virus or firewalls on this machine. I am at a loss on how to fix this so any insight is appreciated.

    Read the article

  • Wireshark Not Displaying Packets From Other Network Devices, Even in Promisc Mode

    - by eb80
    System Setup: 1. MacBook running Mountain Lion. 2. Wireshark installed and capturing packets (I have "capture all in promiscuous mode" checked) 3. I filter out all packets with my source and destination IP using the following filter ("ip.dst != 192.168.1.104 && ip.src != 192.168.1.104") 4. On the same network as the MacBook, I use an Android device (connecting via WiFi) to make HTTP requests. Expected Results: 1. Wireshark running on the MacBook sees the HTTP request from the Android device. Actual Results: 1. I only see SSDP broadcasts from 192.168.1.1 Question: What do I need to do so that Wireshark, like Firesheep, can see and use the packets (particularly HTTP) from other network devices on the same network?

    Read the article

  • Exchange retention work cycle properties

    - by marcwenger
    I've setup retention tags and policies in Exchange, but only run when I execute start-managedfolderassistant Upon running the command get-mailboxserver | fl name,*workcycle*, *ManagedFolderAssistantSchedule*, I noticed the following fields (truncated): ManagedFolderWorkCycle : 1.00:00:00 ManagedFolderWorkCycleCheckpoint : 1.00:00:00 ManagedFolderAssistantSchedule : {Sun.1:00 AM-Sun.9:00 AM, Mon.1:00 AM-Mon.9:00 AM, Tue.1:00 AM-Tue.9:00 AM, Wed.1:00 AM-Wed.9:00 AM, Thu.1:00 AM-Thu.9:00 AM, Fri.1:00 AM-Fri.9:00 AM, Sat.1:00 AM-Sat.9:00 AM} The first two are set to run every day, but does this conflict with what is set in ManagedFolderAssistantSchedule? What's the difference between ManagedFolderWorkCycle and ManagedFolderWorkCycleCheckpoint?

    Read the article

  • Windows 7 - XP Mode - Apache

    - by Howard
    I've setup Virtual PC and XP Mode on my Windows 7 Pro. Using Apache 2.0.52 I have no problems having my website up and running on the Windows 7 machine. But Under VPC/XP Mode the best I can do is Localhost mode. What do I need to do to enable http connections? I need the XP Mode as besides the website I also run a Web BBS and a Dos based (via telnet) BBS. Some of the apps in the Dos BBS just won't work under 64 bit, no matter what setting (capability) are used. Thanks in advance...

    Read the article

  • What is a proper MySql replication configuration for frequent db updates and rare selects?

    - by serg555
    We currently have 1 master db on its own server and slave db on app server. App executes very frequent but light updates (like increasing counters), and occasional (once in a few minutes) heavy selects (which is the most important part of the app). When app was connected only to master db there were no performance issues. With slave db introduction CPU load avg on app server increased to about 6-10 during that heavy select period (from 3-4 as before). When server doesn't run those frequent updates it seems like performance for selects stays within the limits. So I have a feeling that those updates is what is causing the performance drop (also these frequent updates are not critical so if slave db doesn't have them in sync with master for some time it would be ok). What would be a good db replication setup for such kind of app? What are the replication parameters we could tweak? Thanks.

    Read the article

  • Ubuntu NTP issues

    - by Anups
    I am trying to setup the NTP server on Ubuntu machine. Am breaking my head in this particular issue. Getting an error ntpdate[5005]: no server suitable for synchronization found when doing the command ntppdate. Can anyone please help me out in this? /etc/ntp.conf: server 0.ubuntu.pool.ntp.org server 1.ubuntu.pool.ntp.org server 2.ubuntu.pool.ntp.org server 3.ubuntu.pool.ntp.org Also when I gave command netstat -anltp | grep "LISTEN" tcp 0 0 127.0.0.1:53 0.0.0.0:* LISTEN 1816/dnsmasq tcp 0 0 0.0.0.0:22 0.0.0.0:* LISTEN 939/sshd tcp 0 0 127.0.0.1:631 0.0.0.0:* LISTEN 1013/cupsd tcp 0 0 127.0.0.1:39558 0.0.0.0:* LISTEN 5529/rsession tcp 0 0 0.0.0.0:902 0.0.0.0:* LISTEN 1275/vmware-authdla tcp 0 0 127.0.0.1:47304 0.0.0.0:* LISTEN 5822/rsession tcp6 0 0 :::80 :::* LISTEN 1400/apache2 tcp6 0 0 :::22 :::* LISTEN 939/sshd tcp6 0 0 ::1:631 :::* LISTEN 1013/cupsd So what should I do so that it listens on 123? If I get output as PORT STATE SERVICE 123/udp open ntp If I give command nmap -p 123 -sU -P0 192.168.36.198, it means UDP is open right? Then why doesn't it show in the command to to show listening ports?

    Read the article

  • TrueCrypt with blank password

    - by TheSavo
    This might sound like an odd question, but please hear me out. I want to setup my laptop to use TrueCrypt, but without a password. My wife doesn't travel much with her notebook. There are occasions where we will have to travel and bring the notebook with us. This is when we would like the have the encryption. The time it takes to to change the unlock pass-phrase is considerably less than it takes to encrypt the drive. My thoughts are to leave the disk encrypted but with no password. Then when it's time to pack for a trip and bring the laptop, we would then change the pass-phrase and create a rescue disk.

    Read the article

  • WinXP workgroup, 3 routers 3 computers

    - by Silvera
    I have 3 computers with WinXP x86, and 3 Cisco 1800 series routers. I'm trying to create a workgroup so that the 3 computers can share files with eachother. They can ping eachother (without any internet connection), and the routers setup is correctly configured (with interfaces, ip adresses, and ports). But none of the computers can see eachother, even though they are on the same network. My first question would be - can it be done the way it is currently configured - and, if yes, how, or can anyone point me in the right direction?

    Read the article

  • Use HAProxy or Nginx to Load Balance between VPS

    - by xperator
    I want to load balance + failover backup multiple vps webservers hosted on different providers. I heard that for HAProxy you need multiple server under the same subnet, plus a shared (virtual) ip address between load balancers. But it's not possible in my case cause every VPS is on different node/network. Is there a way to use HAProxy in this kind of setup ? ( Please explain how briefly, I don't want to hear your "YES" answer ) What about NginX? Is it possible to achieve same result with Nginx ? (when servers are located on different nets) I know about Round Rubin DNS, but it doesn't provide a real failover solution, neither a load balance between servers.

    Read the article

< Previous Page | 543 544 545 546 547 548 549 550 551 552 553 554  | Next Page >