Search Results

Search found 27800 results on 1112 pages for 'state machine'.

Page 339/1112 | < Previous Page | 335 336 337 338 339 340 341 342 343 344 345 346  | Next Page >

  • Ping reply not getting to LAN machines but getting in Linux router Gateway

    - by Kevin Parker
    I have configured Ubuntu 12.04 as Gateway machine.its having two interfaces eth0 with ip 192.168.122.39(Static) and eth1 connected to modem with ip address 192.168.2.3(through DHCP). ip-forwarding is enabled in router box. Client machine is configured as: ip address 192.168.122.5 and gateway 192.168.122.39 Client machines can ping router box(192.168.122.39).but when pinged 8.8.8.8 reply is not reaching Client machines but in the tcpdump output on gateway i can see echo request for 8.8.8.8 but never echo reply.Is this because of 122.5 not forwarding request to 2.0 network.Can u please help me in fixing this.

    Read the article

  • A Simple Entity Tagger

    - by Elton Stoneman
    In the REST world, ETags are your gateway to performance boosts by letting clients cache responses. In the non-REST world, you may also want to add an ETag to an entity definition inside a traditional service contract – think of a scenario where a consumer persists its own representation of your entity, and wants to keep it in sync. Rather than load every entity by ID and check for changes, the consumer can send in a set of linked IDs and ETags, and you can return only the entities where the current ETag is different from the consumer’s version.  If your entity is a projection from various sources, you may not have a persistent ETag, so you need an efficient way to generate an ETag which is deterministic, so an entity with the same state always generates the same ETag. I have an implementation for a generic ETag generator on GitHub here: EntityTagger code sample. The essence is simple - we get the entity, serialize it and build a hash from the serialized value. Any changes to either the state or the structure of the entity will result in a different hash. To use it, just call SetETag, passing your populated object and a Func<> which acts as an accessor to the ETag property: EntityTagger.SetETag(user, x => x.ETag); The implementation is all in at 80 lines of code, which is all pretty straightforward: var eTagProperty = AsPropertyInfo(eTagPropertyAccessor); var originalETag = eTagProperty.GetValue(entity, null); try { ResetETag(entity, eTagPropertyAccessor); string json; var serializer = new DataContractJsonSerializer(entity.GetType()); using (var stream = new MemoryStream()) { serializer.WriteObject(stream, entity); json = Encoding.UTF8.GetString(stream.GetBuffer(), 0, (int)stream.Length); } var guid = GetDeterministicGuid(json); eTagProperty.SetValue(entity, guid.ToString(), null); //... There are a couple of helper methods to check if the object has changed since the ETag value was last set, and to reset the ETag. This implementation uses JSON to do the serializing rather than XML. Benefit - should be marginally more efficient as your hashing a much smaller serialized string; downside, JSON doesn't include namespaces or class names at the root level, so if you have two classes with the exact same structure but different names, then instances which have the same content will have the same ETag. You may want that behaviour, but change to use the XML DataContractSerializer if you think that will be an issue. If you can persist the ETag somewhere, it will save you server processing to load up the entity, but that will only apply to scenarios where you can reliably invalidate your ETag (e.g. if you control all the entry points where entity contents can be updated, then you can calculate and persist the new ETag with each update).

    Read the article

  • Limit a process's relative (not absolute) processor consumption in Linux

    - by BobBanana
    What is the standard way in Linux to enforce a system policy to limit the relative CPU use of a single process? That is, on a quad-core machine, I never want a process to use more than 2 CPUs at once, even if the process creates more threads. I do not want an absolute time limit, just a relative limit so that one task cannot dominate the machine. This is also different than renice, which allows a process to use all the resources but just politely step aside if others need them too. ulimit is the usual resource limiting tool, but it does not allow such CPU restrictions.. it can limit the number of processes per user, or absolute CPU time, not restrict the maximum number of active threads of a single process. I've found a couple of user-level tools, like CPUlimit, but not a system level tool or setting. Does such a standard resource controller exist in Linux (Red Hat Enterprise, if it matters.) If there is such a limit imposed, how would a user identify it?

    Read the article

  • How do I boot Windows 8 into Safe Mode?

    - by Earlz
    I have Windows 8 Enterprise x64. It's installed a virtual machine. After trying to update VMWare's guest extension it black screened. After resetting it, now the virtual machine won't boot. It just says "automatic repair couldn't repair your PC" Everything I do ends with an error or it coming back to this screen. If I go to Troubleshoot Advanced Options Start up settings and then go to Safe Mode, all it will do is say "preparing automatic repair" and then "diagnosing your PC" and then will come straight back to the "automatic repair couldn't repair your PC" screen. It's as if it doesn't even attempt to go into safe made. Is there anything else I can do before I reformat?

    Read the article

  • Samba share external USB device

    - by bioShark
    I managed to share stuff from my Ubuntu 12.04 to my private network, and the data is visible from a Windows machine. I even shared a hdd that has windows on it. So everything seems to work fine. When I want to share a mounted device (USB pen drive, USB HDD... etc) however, I get from the Windows machine: Access denied on file \... I realize that this is due to the missing rights on the mounted folder. By default a mounted folder gets the equivalent of 700 : drwx------, and the owner myself. But, I can't seem to change the rights on the external device... they remain 700. Is there a special trick I need to do in order to share NTFS mounted usb devices? Thanks P.S. from this Question I see that NTFS devices can not be shared....is this true? It's a bit strange, because I have in my PC 2 HDD's with 3 NTFS partitions, and I can share them without a problem.

    Read the article

  • Generic Windows 98 USB Drive drivers?

    - by Electrons_Ahoy
    I've got an old Win98 machine that I'd like to pull some data off of. (Tens of gigs of data, actually.) The machine still runs fine, but due to the age, it doesn't have a cd burner, and the pre-NT networking won't connect to SMB shares without a fair amount of wrenching on the registry that I don't really want to do. What I really want to do is plug in an external USB hard drive and copy the files. But - Win98 didn't ship with generic USB drive drivers, and external drives don't ship with them these days. Can anyone point me at a reputable generic driver I can install?

    Read the article

  • Windows Server 2012 is telling me it's not activated when it is

    - by LeonidasFett
    I have a Windows Server 2012 Machine were I need to do some development. The machine is a Hyper-V VM. I activated it with a valid key and restarted the VM. After it booted up again, I went into system properties and saw that it said that Windows is activated. However, there is still the message "Activate Windows" on the desktop wallpaper and if try to change something (for example, the look) it tells me that I need to activate first. Is there anybody who has experienced this? The host OS is Windows 8.1. I already installed all updates and tried to activate again using this CMD-Command: slmgr.vbs /ipk {product key, including dashes}

    Read the article

  • How can I make my PHP development environment more efficient?

    - by pixel
    I want to start a home-brew pet project in PHP. I've spent some time in my life developing in PHP and I've always felt it was hard to organize the development environment efficiently. In my previous PHP work, I've used a windows desktop machine and a linux server for development. This configuration had it's advantages: it's easy to configure Apache (and it's modules)/PHP/MySql on a linux box, and, at the time, this configuration was the same like on production server. However, I never successfully set up a debug connection between my Eclipse install and X-debug on server. Transferring files from my local workspace to the server was also very annoying (either ftp or Bazaar script moving files from repository to web root). For my new setup, I'm considering installing everything on my local machine. I'm afraid that it will slow down workstation performance (LAMP + Eclipse), and that compatibility problems will kick-in. What would you recommend? Should I develop using two separate machines? On one? Do you have experience using one of above configurations in your work?

    Read the article

  • How do I turn of "auto-echo" in bash when I 'cd'?

    - by Avery Chan
    I don't know when this started happening but now, every time I cd to a directory it echoes the path right before it changes directories. This happens when I log into a server but doesn't happen on my local machine. The server is running Linux. My local machine is running Mac OS X. I searched the Google as well as looked at the bash man page but I couldn't find anything. My .bashrc/.bash_profile doesn't have anything related to 'cd' (that I know of). How do I modify this "feature"?

    Read the article

  • connecting 2 ubuntu computers as a lan

    - by Brendan Cutjar
    Hi i am trying to connect 2 ubuntu computers as a LAN. In my current setup I have: One machine running ubuntu 11.10 with Realtek Semiconductor Co., Ltd. RTL8101E/RTL8102E PCI Express Fast Ethernet controller While the other machine is running Ubuntu 12.04 with an Atheros Communications Inc AR8152 v1.1 Fast Ethernet The two machines are connected via a switch (Edimax ES-3205P). Unfortunately I still cannot connect the two machines together. Can somebody please show me what to do and how to go about solving this issue?

    Read the article

  • How to add a VM Server to a VM Domain when the host is not on that domain.

    - by Charlie
    I have created A VM running Windows Server 2008 R2, using VMWare. I have configured this as a domain controller running a Windows Server 2008 domain. I have also created another VM running Windows Server 2008 R2. The HOST machine is using Windows 7 Professional 64 bit. When I try to add the second VM into the domain that the first is the DC for it fails as the VM cannot contact the DC. Simple question really. What have I missed? Is it something to do with the configuration on the Host machine? What do I need to do to enable this scenario? Thanks

    Read the article

  • Podcast online sync service

    - by Hannes de Jager
    I download and listen to podcasts on 2 different computers. I would like so somehow sync the metadata such that, when I subscribe to a feed it gets subscribed on my other computer also. When I've downloaded a podcast it gets marked as downloaded on the other machine also. Is there a software + webservice combination that will allow me to do this. That will save its doings to the cloud and update it also? My one machine is Windows and the other one Linux.

    Read the article

  • Google Games Chat #5

    Google Games Chat #5 The Google Games Chat (official motto: "Still not cancelled") is back for yet another rousing debate about industry trends, the state of gaming in general, and, frankly, any other random thoughts that happen to cross our minds. We don't really filter what we say very much. Ask us questions in the moderator link! We might even get around to answering them! From: GoogleDevelopers Views: 0 0 ratings Time: 00:00 More in Gaming

    Read the article

  • Can't access the Internet in VMware Workstation

    - by asunnysunday
    I'm using VMware 7.1.2 in Windows 7 with Ubuntu 11.04 as a guest OS. In the host OS (Windows 7), I can access the Internet without any problems but in the virtual machine I can't access the Internet. I've tried the following but with no success: Use all methods of connecting to the Internet in "Virtual Machine Settings": Bridged, NAT, Custom; none work. Used cabled and wireless connections on the PC - neither of them work. I've used Ubuntu in VMware for several months - previously the Internet was always accessible. Could the cause of this be because I upgraded to Ubuntu 11.04?

    Read the article

  • Send Ctrl+Alt+Del over VNC with Windows Vista and UAC off?

    - by jasonh
    I'm trying to set up a machine here with UltraVNC on Windows Vista with UAC off. The problem is, in this configuration, I can't send Ctrl + Alt + Del (CAD) to the Windows Vista machine, and so I'm stuck at the login screen, which is waiting for the CAD sequence. I'm joined to a domain, so I don't think I can disable the CAD requirement. I can't re-enable UAC either. I also tried using the on-screen keyboard, but it didn't seem to work. Is there a workaround or a solution?

    Read the article

  • TPM had to be reintialized: Does a new recovery password have to be uploaded to AD?

    - by MDMoore313
    Some way some how, a user's machine couldn't get read the bitlocker password off of the TPM chip, and I had to enter the recovery key (stored in AD) to get in. No big deal, but once in the machine, I tried to suspend bitlocker per recovery documentation, and got an error message about the TPM not being initialized. I knew the TPM was on and activated in the BIOS, but Windows still made me reinitialize the TPM chip, and in the process it created a new TPM owner password. I found that odd because it prompted me to save this password or print it (there wasn't an option not to), but it made no reference of a recovery password, nor did it back this password up to AD. After the user took her laptop and left I started thinking that if the TPM password change, does the recovery password change also? If so, that new recovery password will need to be uploaded to AD, but MS' documentation doesn't make that clear, and doesn't back up the new recovery key (if one exists) to AD automatically when the group policy says it must, and from a network standpoint AD is accessible.

    Read the article

  • Do you leave Windows Automatic Updates enabled on your production IIS server?

    - by Nobody
    If you were running a 24/7 website on Windows Server 2003 (IIS6). Would you leave the Windows automatic update feature enabled or would you turn it off? When enabled, you always get the latest security patches and bug fixes automatically as soon as they're available, which is the most secure choice. However, the machine will sometimes get automatically rebooted to apply the updates leading to a couple of minutes of downtime in the middle of the night. Also, I've seen rare occasions where the machine does not restart correctly resulting in further downtime. If auto updates are off, when do you apply the patches? I guess you have to use a load balancer with multiple web servers and rotate them out of the production site, apply patches manually, and put them back in. This can be logistically inconvenient when the load balancer is managed by a hosting company. You will also have machines in production that don't always have the latest security patches and you have to routinely spend time deciding which patches to apply and when.

    Read the article

  • Solutions for "Maintenance Mode"

    - by Ka Lyse
    Given a web application running across 10+ servers, what techniques have you put in place for doing things like altering the state of your website so that you can implement certain features. For instance, you might want to: Restrict Logins/Disable Certain Features Turn Site to "Read Only" Turn Site to Single "Maintenance Mode" page. Doing any of the above is pretty trivial. You can throw a particular "flag" in an .ini file, or add a row/value to a site_options table in your database and just read that value and do the appropriate thing. But these solutions have their problems. Disadvantages/Problems For instance, if you use a file for your application, and you want to switch off a certain feature temporarily, then you need to update this file on all servers. So then you might want to look at running something like ZooKeeper, but you are probably overcomplicating things. So then, you might decide that you want to store these "feature" flags in a database. But then you are obviously adding unncessary queries to each page request. So you think to yourself, that you will throw memcached in to the mix and just cache the query. Then you just retrieve all of your "Features" from memcached and add a 2ms~ latency to your application on every page. So to get around this, you decide to use a two tier-cache system, whereby you use an inmemory cache on each machine, (like Apc/Redis etc). This would work, but then it gets complicated, because you would have to set the key/hash life to perhaps 60 seconds, so that when you purge/invalidate the memcached object storing your "Features" result, your on machine cache is prompt enough to get the the new states. What suggestions might you have? Keeping in mind that optimization/efficiency is the priority here.

    Read the article

  • Linux PHP web server horribly slow when accessed from any windows browser

    - by Ed Harcourt
    I have a Linux server (Ubuntu 10.04) running apache2 and PHP. Everything runs fine when accessing a page from any browser from another Linux machine or Mac. But when I try to access a page from any combination of Windows machine and browser I get about a 30 second delay before the page comes back. Accessing a plain old HTML file from the Windows browser runs lickity split. So it seems to be just PHP. MySQL is installed but a simple test page that uses no MySQL is still slow. I don't think it is DNS because if I hard code the IP address in the URL nothing changes. There doesn't seem to be anything in the log files that I can tell. What could be causing this behavior on Windows clients?

    Read the article

  • Error 5A - Internal CPU Error

    - by Mitch
    So I built myself a pc for the first time. The machine has been running properly, and without issue, since March of this year. Today after an hour or so of gaming the machine shutdown without warning. Any attempts to reboot have been unsuccessful. I have managed to get it to post, twice, but it doesn't get as far as the OS before it shuts off again. I pulled the second video card so I could see the error code LED and it appears to be showing 5A which matches to "Internal CPU Error". Am I looking at a failed CPU or is this a symptom of another issue? Any feedback you can provide is most appreciated. Here's my parts list: CPU: Intel Core i7-4930K 3.4GHz 6-Core CPU Cooler: Corsair H105 73.0 CFM Liquid Motherboard: Asus X79 Deluxe ATX LGA2011 Ram: G.Skill Ripjaws X Series 32GB (4 x 8GB) DDR3-1600 Video card (x2): EVGA GeForce GTX 770 2GB Power supply: Corsair 1000W ATX12V / EPS12V

    Read the article

  • Dual boot centOS and Win7

    - by user1855965
    I posted this on stackoverflow, but it looks like superuser would be more appropriate. I have a CentOS 5 machine that runs Windows 7 as a dual boot. CentOS is the main OS and each OS is set up in a specific hard drive. This was set up before I joined the company and I don't really have need to run Windows now. My question is: can I, from CentOS, reformat the Windows HD, change GRUB settings and get the HD to be available on CentOS? Happy to provide more info if this helps. Many thanks for your help and apologies if this is a very simple issue... I don't want to blindly test things on this machine as it is used on a daily basis by several users.

    Read the article

  • Linux - Network Sharing a local NTFS usb drive

    - by Jonathan Rioux
    I have an external hard drive formated in NTFS which I would like to be able to access by the network. I want to make a network share out of it. I also have a Debian machine running in my house and I then got an idea. I want to plug in my external hard drive (usb) into my Debian machine, and make a windows share with it, maybe with Samba, so I will be able to access it from my Windows 7 laptop and see it as a network share. Additionally, how can I restrict specific folders of that network share, and allow only specific folders to specific users? For instance, I would like to give my girlfriend access to a folder of her name so she can put her files and so she wont be able to see the stuff in my folder...

    Read the article

  • can't run binaries or shell scripts

    - by hyperboreean
    I am running Debian testing and I am not able to run any binary or shell script. I keep getting "No such file or directory". The umask is the default one and I haven't fooled around with the paths. Also, I am aware of this question, but it doesn't work out for me - I compiled my code on this machine and trying to run it on the same machine. Also, all of my shell scripts have the correct shebang. Any advices?

    Read the article

  • Should we install the OS on an SSD or not when running virtual machines?

    - by Raghu Dodda
    I have a new Dell Mobile Precision M6500 laptop with 8 GB RAM. it has two hard drives - 500 GB @7200 RPM and a 128 GB SSD. The main purpose of these laptop is software development in virtual machines. The plan is to install the base OS (Windows 7) and all the programs in the 500 GB drive, and let the SSD only contain the virtual machine images. It is my understanding that the we get most performance from the virtual machines if the images are on a separate hard drive than the base OS. Is this the way to go, or should I install the OS on the SSD as well? What are the pros and cons? The virtual machine images would be between 20 - 30 GB, and I might run 1 or 2 at a time.

    Read the article

< Previous Page | 335 336 337 338 339 340 341 342 343 344 345 346  | Next Page >