Search Results

Search found 17749 results on 710 pages for 'connection pool'.

Page 392/710 | < Previous Page | 388 389 390 391 392 393 394 395 396 397 398 399  | Next Page >

  • Licensing approach for .NET library that might be used desktop / web-service / cloud environment

    - by Bobrovsky
    I am looking for advice how to architect licensing for a .NET library. I am not asking for tool/service recommendations or something like that. My library can be used in a regular desktop application, in an ASP.NET solution. And now Azure services come into play. Currently, for desktop applications the library checks if the application and company names from the version history are the same as the names the key was generated for. In other cases the library compares hardware IDs. Now there are problems: an Azure-enabled web-application can be run on different hardware each time (AFAIK) sometimes the hardware ID for the same hardware changes unexpectedly checking the hardware ID or version info might not be allowed in some circumstances (shared hosting for example) So, I am thinking about what approach I can take to architect a licensing scheme that: is friendly to customers (I do not try to fight piracy, but I do want to warn the customer if he uses the library on more servers than he paid for) can be used when there is no internet connection can be used on shared hosting What would you recommend?

    Read the article

  • Ubuntuone fails to sync with 'File Sync starting...' displayed

    - by a different ben
    I am on 12.04 using ubuntuone-client 3.0.1-0ubuntu1.0.1. I actually have two machines that I sync with, having the same Ubuntu version and ubuntuone-client version. One is fine, the other is not. File sync has frozen within a user-defined folder under my home folder. The graphical client reports in the top-right corner: 'File Sync starting...', but this doesn't change. I have two files with changes that show a syncing overlay in Nautilus. They are both very small text files. Here are some details: harb@joan:~$ u1sdtool --status State: READY connection: With User Not Network description: ready to connect is_connected: False is_error: False is_online: False queues: WORKING harb@joan:~$ u1sdtool --current-transfers Current uploads: 0 Current downloads: 0 The status seems to suggest that I am not connected to a network, however I am connected to a network - in fact I am accessing this machine via NX. Is it not working because I am connected via NX? Happy to provide other info, just not sure what would be useful.

    Read the article

  • 127.0.0.1:9051 doesnt work after apache, mysql, php installation?

    - by Rana Muhammad Waqas
    I have installed apache2, mysql, and php and now it doesnt let Vidalia run on localhost. i tried to change the TCP connection (controlport) to any other ip 192.168.0.40 and tried to change the default port 9051 to any other but that doesnt work. I thought apache is running so i used this command sudo service apache2 stop but that still doesnt work. So now when i type 127.0.0.1:9051 in browser it says and if i type only type 127.0.0.1 after stopping the apapche2 service with the command mentioned above it says unable to connect I am not sure what to do now Help!

    Read the article

  • Using virtualization infrastructure for J2EE application distribution- viable alternative?

    - by Dan
    Our company builds custom J2EE web solutions. At the moment, we use standard J2EE distribution mechanisms (ear/war archives). Application servers are generally administered by our clients' IT departments and since we do not have complete control over the environment, a lot of entropy can be introduced into the solution. For example: latest app. server patch not applied conflicting third party libraries inside the app. server root server runtime and tuning parameters not configured (for example, number of connections in database pool) We are looking into using virtualization infrastructure for J2EE application distribution. Instead of sending the ear/war archive, we’d send image with application server node and our application preinstalled. Some of the benefits are same as using with using virtualization infrastructure in general, namely better use of hardware resources. For us, it reduces the entropy of hosting infrastructure - distributing VM should be less affected by hosting environment. So far, the downside I see can be in application server licenses, here they will have to use dedicated servers for our solution, but this is generally already done that way. Also, there is a complexity with maintaining virtualization infrastructure, but this is often something IT departments have more experience with than with administering and fine-tuning J2EE solutions. Anyone has experience with this model? What are the downsides? Will we not just replace one type of complexity with other?

    Read the article

  • AutoSSH for a robust tunnel

    - by Budric
    I'm trying to start an ssh tunnel from A to B and have it run despite things like: period network/wifi drops on A and remote server reboot on B. My ssh tunnel starts using upstart script on A with event start on (net-device-up IFACE=eth0) I've found autossh which is supposed to handle these kinds of things, but had some trouble getting it to work. The upstart executes: autossh -M 0 -2qTN -o "ServerAliveInterval 30" -o "ServerAliveCountMax 2" -L 5678:somehost:5678 user@B However when I log into B and kill -9 that tunnel session, autossh just exits with "Connection to B closed by remote host." That's not what I expected autossh to do. Any advice on how to set this up? Any GUI service monitoring utilities out there that essentially display a green light if a service is up? Thanks.

    Read the article

  • Can we set up svn server on a local computer without any network access?

    - by Aitezaz Abdullah
    I want to set up an SVN repository on my computer without any network access. I am working on a code without any collaborator, so I don't want it to be publicly available. I read the following post. http://stackoverflow.com/questions/6001445/local-source-control-repository-cross-platform but this post suggests using online svn repository services that give free repositories. In that case, my code will be publicly available (as is included in the terms of free plans). So I was wondering if I can set up a local server on my windows xp machine that only I access even when I don't have any internet connection?

    Read the article

  • Did Blowing Into Nintendo Cartridges Really Help?

    - by Jason Fitzpatrick
    Anyone old enough to remember playing cartridge-based games like those that came with the Nintendo Entertainment System or its successors certainly remembers how blowing across the cartridge opening always seemed to help a stubborn game load–but did blowing on them really help? Mental Floss shares the results of their fact finding mission, a mission that included researching the connection mechanism in the NES, talking to Frank Viturello (who conducted an informal study on the effects of moisture on cartridge connectors), and otherwise delving into the history of the phenomenon. The most interesting part of the analysis, by far, is their explanation of how blowing on the cartridge didn’t do anything but the ritual of removing the cartridge to blow on it did. Hit up the link below for the full story. Did Blowing into Nintendo Cartridges Really Help? [Mental Floss] How Hackers Can Disguise Malicious Programs With Fake File Extensions Can Dust Actually Damage My Computer? What To Do If You Get a Virus on Your Computer

    Read the article

  • HDMI audio output problems with Radeon card

    - by Matt Robinson
    No matter how much advice I follow, I still cannot get any audio to come out through my HDMI connection. I've tried downloading the latest proprietary FGLRX graphics drivers, and I've also gone into /etc/default/grub and altered GRUB_CMDLINE_LINUX_DEFAULT="quiet splash" into this GRUB_CMDLINE_LINUX_DEFAULT="quiet splash radeon.audio=1" But still I cannot get any sound nor can I get HDMI sound output to show up in Sound Settings or Pulse Audio Volume Control. I can get video to show up just fine on my monitor through HDMI so I'm sure this problem is fixable! I know this is an old problem with Ubuntu, but any new strategies into the problem would be much appreciated. With that being said, here are some of my specs: Ubuntu 12.04 LTS AMD Radeon HD 7520G

    Read the article

  • Reproducing a Conversion Deadlock

    - by Alexander Kuznetsov
    Even if two processes compete on only one resource, they still can embrace in a deadlock. The following scripts reproduce such a scenario. In one tab, run this: CREATE TABLE dbo.Test ( i INT ) ; GO INSERT INTO dbo.Test ( i ) VALUES ( 1 ) ; GO SET TRANSACTION ISOLATION LEVEL SERIALIZABLE ; BEGIN TRAN SELECT i FROM dbo.Test ; --UPDATE dbo.Test SET i=2 ; After this script has completed, we have an outstanding transaction holding a shared lock. In another tab, let us have that another connection have...(read more)

    Read the article

  • set proxy for vpn server on ubuntu server 12.4

    - by Morteza Soltanabadiyan
    I have a vpn server with HTTPS, L2TP , OPENVPN , PPTP. i want to set proxy in the server so all connection that comes from vpn clients use the proxy that i set in my server. I made a bash script file for it , but proxy not working. gsettings set org.gnome.system.proxy mode 'manual' gsettings set org.gnome.system.proxy.http enabled true gsettings set org.gnome.system.proxy.http host 'cproxy.anadolu.edu.tr' gsettings set org.gnome.system.proxy.http port 8080 gsettings set org.gnome.system.proxy.http authentication-user 'admin' gsettings set org.gnome.system.proxy.http authentication-password 'admin' gsettings set org.gnome.system.proxy use-same-proxy true export http_proxy=http://admin:[email protected]:8080 export https_proxy=http://admin:[email protected]:8080 export HTTP_PROXY=http://admin:[email protected]:8080 export HTTPS_PROXY=http://admin:[email protected]:8080 Now , I don't know what to do to make a global proxy for server and all vpn clients use it automatically.

    Read the article

  • Cannot connect to a Wifi network chipset AR9287

    - by Fritz
    After updating my newly installed ubuntu 10.10. I can't connect to the wifi network anymore although it is showing it. Tried to install manufacturer driver INF file via ndiswrapper only makes things worst - lost my wifi. Downloaded and installed the compat-wireless driver for the Atheros AR9287. After installing, i got my wifi connection again but after rebooting I can't reconnect again. The Wifi access point is showing 2/5 to 3/5 signal strength. Laptop: Acer Aspire 4741G Wifi Adapter: Atheros AR9287 OS: Ubuntu 10.10 64bit

    Read the article

  • Free tiered storage automation in linux?

    - by NginUS
    I have a couple virtualized fileservers running in QEMU/KVM on ProxmoxVE. The physical host has 4 storage tiers with significant performance variances. They're attached both locally and via NFS. These will be provided to the fileserver(s) as local disks, abstracted into pools, and handling multiple streams of data for the network. My aim is for this abstraction layer to intelligently pool the tiers. There's a similar post on the site here: Home-brew automatic tiered storage solutions with Linux? (Memory - SSD - HDD - remote storage) in which the accepted answer was a suggestion to abandon a linux solution for NexentaStor. I like the idea of running NexentaStor. It almost fits the bill. NexentaStor provides Hybrid Storage Pools, and I love the idea of checksumming. 16TB without incurring licensing fees is a huge plus as well. After the expense of the hardware, free is about all my budget can handle. I don't know if zfs pools are adaptive or dynamically allocated based on load, but it becomes irrelevant since NexentaStor doesn't support virtio network or block drivers, which is a must in my environment. Then I saw a commercial solution called SmartMove: http://www.enigmadata.com/smartmove.html And it looks like a step in the right direction, but I'm so broke I'd be wasting their time to even ask for a quote, so I'm looking for another option. I'm after a linux implementation that supports virtio drivers, and I'm at a loss as to which software is up to it.

    Read the article

  • HTG Explains: What is the Windows Page File and Should You Disable It?

    - by Chris Hoffman
    Windows uses a page file to store data that can’t be held by your computer’s random-access memory when it fills up. While you can tweak the page file settings, Windows can manage the page file fine on its own. The Windows page file is somewhat misunderstood. People see it as the cause of slowdowns because it’s slower to use the page file than your computer’s RAM, but having a page file is better than not having one. Image Credit: Blake Patterson on Flickr HTG Explains: What is the Windows Page File and Should You Disable It? How To Get a Better Wireless Signal and Reduce Wireless Network Interference How To Troubleshoot Internet Connection Problems

    Read the article

  • Atheros 9250 wireless card keeps on disconnecting

    - by Vizi Balázs
    So I bought a Lenovo G575 laptop for learning purposes and found that the connection just simply disconnects if i'ts in use for downloading and it also does disconnect in every 2-5 minutes while browsing.I tested the card with Win7, it worked nicely.Since then I searched through all the topics connected to this card, also tried using older distros, Maverick ended up with the same results as the actual 12.04 distribution.I dont think it's the router cause i could access it from Win7.It's causing me a trouble using an external usb, cause it's taken from another PC in the house.I would be really grateful if someone could explain what should i do to fix it. Thanks in advance for the help

    Read the article

  • Web framework able to handle many concurrent users [closed]

    - by Jonas
    Social networking sites needs to handle many concurrent users e.g. for chat functionality. What web frameworks scales well and are able to handle more than 10.000 concurrent users connected with Comet or WebSockets. The server is a Linux VPS with limited memory, e.g. 1GB-8GB. I have been looking for some Java frameworks but they consume much memory per connection. So I'm looking for other alternatives too. Are there any good frameworks that are able to handle more than 10.000 concurrent users with limited memory resources?

    Read the article

  • How to configure IIS 7.5 to allow special chars in Url for ASP.NET 3.5?

    - by Sebastian P.R. Gingter
    I'm trying to configure my IIS 7.5 to allow specials chars in the url for ASP.NET. This is important to support wide-spread legacy url's on a new system. Sample url: http://mydomain.com/FileWith%inTheName.html This would be encoded in the url and requested as http://mydomain.com/FileWith25%inTheName.html This simply works, when creating a new web in IIS 7.5, placing a file with the percentage sign in the file name in the web root and pointing the browser to it. This does not work, however, when the web site is an ASP.NET application. ASP.NET always returns a 400.0 - Bad Request error in the WindowsAuthentication module from the StaticFile handler, when pointing to that url. It however displays the requested url correctly and also resolves correctly to the correct physical file (the information from the field 'Physical Path' from the Server error page points to the physically available file). There are hints on how to enable this, so I followed the instructions on these websites step by step: http://dirk.net/2008/06/09/ampersand-the-request-url-in-iis7/ http://adorr.net/2010/01/configure-iis-to-accept-url-with-special-characters.html The second one actually sums up the information from the first post and adds some more information about x64 systems (we're running x64) and on an additional web.config change for this. I tried all that, and still can't get this running from an asp.net web application. And yes: I rebooted after applying the registry changes. So, what do I have to do in addition to the settings described in above posts, to support the legacy url's which contain percentage characters? Additional info: Application Pool mode is integrated. Push after some days. No idea anyone?

    Read the article

  • Do programmers possess the means of production?

    - by Peter Turner
    I was listening to The Servile State by Hilare Belloc this morning and pondering whether or not I possessed the means of production, as did the peasant of the middle ages; as did not his descendants after the oligarchs of England forced him into servility. The means of production was the arable land that the serf was seated on, which even though not legally his, was illegal to evict him from. So, as programmers, with the hither-to-unknown supply of free tools and resources, have we reclaimed as a class of workers, unlike any others, the means of production. Given the chance, a midrange PC and a stable internet connection, could we not each of us be wholly self sufficient and not just wage earners?

    Read the article

  • Problem upgrading from 13.04 to 13.10

    - by Charles
    Part way through upgrading from 13.04 to 13.10 the process ground to a halt with an error message. Now on retrying by going to 'Check for updates' I get the following: Failed to load the package list This is a serious problem. Try again later. If this problem appears again, please report an error to the developers. E:Encountered a section with no Package: header, E:Problem with MergeList /var/lib/apt/lists/archive.ubuntu.com_ubuntu_dists_saucy_universe_i18n_Translation-en%%5fGB, E:The package lists or status file could not be parsed or opened. Problem reported but my question is, "what can I do now?; Do I have to do a fresh install?; if so will settings etc. in my Home folder (on its own partition) be saved?" 13.04 still seems to be working perfectly, while upgrading I had a terrible internet connection varying between 'dead slow' and 'dead stop', not sure if that caused the problem.

    Read the article

  • Unable to Upgrade from 13.04 to 13.10

    - by Mohit
    Following is what I get as error Error during update A problem occurred during the update. This is usually some sort of network problem, please check your network connection and retry. W:Failed to fetch http://bg.archive.ubuntu.com/ubuntu/dists/natty-backports/main/source/Sources 404 Not Found , W:Failed to fetch http://bg.archive.ubuntu.com/ubuntu/dists/natty-backports/restricted/source/Sources 404 Not Found , W:Failed to fetch http://bg.archive.ubuntu.com/ubuntu/dists/natty-backports/universe/source/Sources 404 Not Found , W:Failed to fetch http://bg.archive.ubuntu.com/ubuntu/dists/natty-backports/multiverse/source/Sources 404 Not Found , E:Some index files failed to download. They have been ignored, or old ones used instead. Restoring original system state I am using command line "do-release-upgrade" I have changed my dns to google public dns and verified that using nslookup and host -v could not find bg.archive.ubuntu.com in any of the source list files. not only in that in another tab I have ping google running just to verify that network is not down. Network has no issues.

    Read the article

  • Download mirrors down after fresh installation

    - by user169866
    I dont know whats the problem but it seems after every fresh installation of ubuntu in any device (have started noticing this problem since ubuntu 11.04) the connection to the software sources are down. I cannot download any software from either apt-get or software center. Also i cannot use the Universal Source for packages like chromium, etc. And then suddenly after a few days, the mirrors are back up again!! Its normal back again! I dont know what the problem is, but it happens every time. Is there any solution to this or we just have to wait for a few days till the mirrors are back again? (and I have also tried changing the download server, It doesn't helps!)

    Read the article

  • Broadcom BCM4313 wireless slow and high-latency

    - by Florin Andrei
    Ubuntu 12.10 64 bit on a Dell Latitude E6330 laptop. Wireless is pretty slow. It gets connected quick enough, but then it acts like a dialup connection. My ssh sessions over WiFi are slow and laggy. Even browsing is slow, the pages are loading like it's 1998. This does not depend on the access point, it's the same both at home and at work. Other systems work fine on these access points. I had an older Dell laptop before, different WiFi hardware, and it was much faster over the same wireless access points. Is this a known issue with this hardware? If so, any solutions?

    Read the article

  • Warning about SSL ceritificate, am I under attack ?

    - by Bunny Rabbit
    Lately I've been getting a lot of warnings about SSL certifications on my pc, Empathy keeps telling me that Facebook's certificate is self-signed and can't be trusted, and also, there are occasional warnings in Google-Chrome about security. I remember the last one saying that that the page is secured but some of the resources that the page is using are not from a secure connection, something like that. Is my pc hacked / under attack? How can I check that, and if so, how can I safeguard myself? PS: One thing that comes to my mind is that I might be under an arp poisoning / spoofing attack.

    Read the article

  • Hybrid wireless network repeating

    - by Oli
    Summary: I'd like to use two Ubuntu computers to extend/compliment an existing wireless access point. I have a network which currently looks a bit like this: What the diagram doesn't show is the interference caused by our house. It's a wifi-blocking robot sent here from the past. The two wired computers are in areas where the signal is most blocked (not by design, just a happy co-incidence). Both wired computers have fairly good network cards. They're both Ubuntu machines and I would like to turn them into additional base stations. I know I could throw more networking hardware at this (network extenders or cable in additional, pure wireless access points) but I've got two Linux machines sitting in ideal places and I feel like they should be able to help me out. I've tried ad-hoc networks but I need something that is a lot more transparent (eg you can migrate from base to base without a connection dropping); it should look like one network to clients.

    Read the article

  • GConf Error: No D-BUS daemon running?! How to reinstall or fix?

    - by v2r
    After installing Konquerer and restarting my laptop i got the following error while trying to open, edit or access files as root from within Terminal (which is existencial for me). root@linuxBox:/home/v2r# gnome-open /home/ (gnome-open:2686): GConf-WARNING **: Client failed to connect to the D-BUS daemon: //bin/dbus-launch terminated abnormally with the following error: No protocol specified Autolaunch error: X11 initialization failed. GConf Error: No D-BUS daemon running root@linuxBox:/home/v2r# No protocol specified Could not parse arguments: Cannot open display: Also it seems, that dbus is not installed properly anymore in /bin/ and /usr/bin/ See screenshot: How would i go about fixing this problem and thank you in advance?!!! Thank you for your answer SirCharlo! It does not resolve the problem at all. Please note, that it only happens while beeing root! root@linuxBox:/home/v2r# gnome-open /home/ (gnome-open:5170): GConf-WARNING **: Client failed to connect to the D-BUS daemon: Failed to connect to socket /tmp/dbus-2RdCUjrZ9k: Connection refused GConf Error: No D-BUS daemon running root@linuxBox:/home/v2r# No protocol specified Could not parse arguments: Cannot open display:

    Read the article

  • Getting Ethernet driver to crippled machine

    - by Richard
    The reply at How do I install drivers for the Atheros AR8161 Ethernet controller? probably helps me. I have a Toshiba S855, bought yesterday. lspci -vv | grep Atheros command from the terminal shows me I have the AR8161 Gigabit Ethernet controller (rev 10) However the answer sudo apt-get install ... requires an internet connection. How can I, using another machine, capture the necessary information on that other machine to copy across to the crippled machine, and then how do I build from there. Thanks in anticipation

    Read the article

< Previous Page | 388 389 390 391 392 393 394 395 396 397 398 399  | Next Page >