Search Results

Search found 23265 results on 931 pages for 'justin case'.

Page 466/931 | < Previous Page | 462 463 464 465 466 467 468 469 470 471 472 473  | Next Page >

  • Route using certain IP address

    - by spa
    I have a server with two public IPs. Both IPs are added to eth0 using ip addr add. Now I'd like to contact a server which uses IP address filtering. Only requests are allowed which use the second IP address. Is there are way to set this up using the standard route command in Linux? I guess that's not the case. So the only solution I see right now: Setup a virtual device let's say eth0:0 and bind the second IP address to it. Then I can reference the device in the route command. Edit: I can't use the second IP as primary one easily as this IP is used as failover IP.

    Read the article

  • How can we add arch specific conflicts tag when building .deb package?

    - by Sphopale
    We are trying to build multi-arch supported i386 .deb package. There are two .deb packages build on i386 X1 & X2 (X2 is a subset of X1 binaries). X1 <- X2 conflict each other when installing . Only one .deb package can be installed at any instance. We similarly have binaries on xa64 arch. Again on xa64, there are two .deb packages X1 & X2 (X2 is a subset of X1 binaries). X1 <- X2 conflict each other when installing . Only one .deb package can be installed at any instance. In case of multi-arch i386 .deb package,i386 .deb packages (X1 & X2) can be installed on xa64 along side with 64bit (X1 & X2) However I see that when installing X1:i386 & X1:amd64 can co-exist However, it throws conflict error when trying to install X1:i386 & X2:amd64 In short, Can we mark package to conflict based on arch Conflict: X2:i386 X1:i386 package should only conflict with X2:i386 & allow other packages to co-exist X1:amd64 package should only conflict with X2:amd64 & allow other packages to co-exist X1:i386 can co-exist with X1:amd64 OR X2:amd64 X2:i386 can co-exist with X1:amd64 OR X2:amd64 Thanks for your reply

    Read the article

  • Cannot dump svn repository

    - by vinga
    I've a problem with my svn repo. I cannot use it, I even cannot dump it. svnadmin verify repo returns Can't set position pointer in file 'svn/db/revs/0/0' When I try to dump repo (no matter what revision range), console output shows: * Dumped revision 0. svnadmin: Final line in revision file missing space I've googled that this may be connected with wrong version apr apache2 library, but I have other repositories which work good, so I thing this isn't the case. Is there any way to save at least some files from my repo? Can svn repo get corrupted so easily (probably after power-cut, however I'm not sure).

    Read the article

  • Eclipse style autohotkey commands

    - by Ph4g3
    Is autohotkey capable of interpreting hotkeys in the style of Ctrl + Shift + W, W? I would assume a script like the following would work: ^+W:: ; Windows hotkeys (Ctrl + Shift + W) O:: Run Outlook ; Subsequent 'o' pressed => Run outlook E:: Run Explorer ; return From the documentation I note that these are called vertically stacked hotkeys and cause each line to perform the same action. In the case above, I think Ctrl+Shift+W and o will both cause outlook to be launched, whereas pressing e would cause explorer to be launched. What I would like is Ctrl+Shift+W, O runs outlook and Ctrl+Shift+W, e runs windows explorer. Is there any way to cause a hotkey to perform context specific actions (much like in Eclipse), where I can press Ctrl+Shift+W to activate a block of specific hotkeys?

    Read the article

  • What are the pros & cons of these MySQL engines for OLTP -- XtraDB, PBXT, or TokuDB?

    - by Continuation
    I'm working on a social website with an approximate read/write split of 90/10. Trying to decide on a MySQL engine. The ones I'm interested in are: XtraDB PBXT TokuDB What are the pros and cons of them for my use case? A few specific questions: PBXT uses log-based structure that avoids double-writes. It sounds very elegant, but the benchmark I've seen doesn't show any/much advantages over XtraDB. Do you have any experience with PBXT/XtraDB you can share? TokuDB sounds VERY interesting. But all the benchmarks I've seen are about single-threaded bulk inserts - inserting 100M rows for example. that's not very relevant for OLTP. What about its performance with large number of concurrent threads writing and reading at the same time? Anyone has tried that?

    Read the article

  • Using mod_rewrite for a Virtual Filesystem vs. Real Filesystem

    - by philtune
    I started working in a department that uses a CMS in which the entire "filesystem" is like this: create a named file or folder - this file is given a unique node (ex. 2345) as well as a default "filename" (ex. /WelcomeToOurProductsPage) and apply a template assign one or more aliases to the file for a URL redirect (ex. /home-page-products - can also be accessed by /home-page-products.aspx) A new Rewrite command is written on the .htaccess file for each and every alias Server accesses either /WelcomeToOurProductsPage or /home-page-products and redirects to something like /template.aspx?tmp=2&node=2345 (here I'm guessing what it does - I only have front-end access for now - but I have enough clues to strongly assume) Node 2345 grabs content stored in a SQL Db and applies it to the template. Note: There are no actual files being created on the filesystem. It's entirely virtual. This is probably a very common thing, but since I have never run across this kind of system before two months ago, I wanted to explain it in case it isn't common. I'm not a fan at all of ASP or closed-sourced systems, so it may be that this is common practice for ASP developers. My question, that has taken far too long to ask, is: what are the benefits of this kind of system, as opposed to creating an actual file hierarchy? Are there any drawbacks to having every single file server call redirected? To having the .htaccess file hold rewrite rules for every single alias?

    Read the article

  • Best way to administer a website remotely

    - by Mark Szymanski
    I have a Windows computer running an intranet website with IIS and I was wondering what the best way was to administer it from another computer, in this case, a Mac. What I want to do: Be able to edit pages from my Mac. VNC into the server because it is 'headless'. (Already have this set up) My current file syncing setup: I have Dropbox setup to sync files between the computers and then use PureSync to sync the files in the Dropbox folder into the wwwroot folder. Is there a better (faster/easier) way I could do all this? Thanks in advance!

    Read the article

  • Apply rewrite rule for all but all the files (recursive) in a subdirectory?

    - by user784637
    I have an .htaccess file in the root of the website that looks like this RewriteRule ^some-blog-post-title/ http://website/read/flowers/a-new-title-for-this-post/ [R=301,L] RewriteRule ^some-blog-post-title2/ http://website/read/flowers/a-new-title-for-this-post2/ [R=301,L] <IfModule mod_rewrite.c> RewriteEngine On ## Redirects for all pages except for files in wp-content to website/read RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteCond %{REQUEST_URI} !/wp-content RewriteRule ^(.*)$ http://website/read/$1 [L,QSA] #RewriteRule ^http://website/read [R=301,L] RewriteBase / RewriteRule ^index\.php$ - [L] RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteRule . /index.php [L] </IfModule> My intent is to redirect people to the new blog post location if they propose one of those special blog posts. If that's not the case then they should be redirected to http://website.com/read. Nothing from http://website.com/wp-content/* should be redirected. So far conditions 1 and 3 are being met. How can I meet condition 2?

    Read the article

  • Linux, PHP, and mysqlnd?

    - by bob-the-destroyer
    Looks like most common distros build their PHP packages without mysqlnd, instead relying on the old libmysql library. Are there any distros that do include the new mysql native driver even as an option in any of their officially supported repositories? If not, is it possible (and if so, how) to rebuild these three (mysql, mysqli, & pdo_mysql) extensions without having to rebuild the entire PHP install? In my case, there are several PHP functions I need that are available only if these Mysql extenstions are built against mysqlnd. So I'd be fine installing a new OS to get these if the OS offers and supports it.

    Read the article

  • Implementing Explosions

    - by Xkynar
    I want to add explosions to my 2D game, but im having a hard time with the architecture. Several game elements might be responsible for explosions, like, lets say, explosive barrels and bullets (and there might be chain reactions with close barrels). The only options i can come up with are: 1 - Having an array of explosions and treat them as a game element as important as any other Pros: Having a single array which is updated and drawn with all the other game element arrays makes it more organized and simple to update, and the explosive barrels at a first glance would be easy to create, simply by passing the explosion array as a pointer to each explosive barrel constructor Cons: It might be hard for the bullets to add an explosion to the vector, since bullets are shot by a Weapon class which is located in every mob, so lets say, if i create a new enemy and add it to the enemy array, that enemy will have a weapon and functions to be able to use it, and if i want the weapon (rocket launcher in this case) to have access to the explosions array to be able to add a new one, id have to pass the explosion array as a pointer to the enemy, which would then pass it to the weapon, which would pass it to the bullets (ugly chain). Another problem I can think of is a little more weird: If im checking the collisions between explosions and barrels (so i create a chain reaction) and i detect an explosion colliding with a barrel, if i add a new explosion while im iterating the explosions java will trow an exception. So this is kinda annoying, i cant iterate through the explosions and add a new explosion, i must do it in another way... The other way which isnt really well thought yet is to just add an explosive component to every element that might explode so that when it dies, it explodes or something, but i dont have good ways on implementing this theory either Honestly i dont like either the solutions so id like to know how is it usually done by actual game developers, sorry if my problem seems trivial and dumb.

    Read the article

  • How many xml http requests is too much for a pc to handle?

    - by Uri
    I'm running mediawiki on an apache on a regular pc running vista (don't know the specific specs, but I'm guessing at least duo core 2 2 giga hertz processor, broadband connection (500 kb/s at least, probably 1 mega). I want to use the MediaWiki api to send a lot of requests to this server. Most of the time the requests will be sent through LAN (but sometimes through the internet). I'm talking thousands of requests every few seconds at worst case. (A lot of these requests may repeat themselves, I guess some sort of cache would help) Will the server handle this, or do I need a stronger/dedicated computer? (I'm not looking for specific yes/no, but just want to get an idea as to what configuration of computer will support how many request per second) Thanks

    Read the article

  • How does Linux's unlink on a NTFS filesystem differs from Window's own implementation?

    - by DavideRossi
    I have an external USB disk with an NTFS filesystem on it. If I remove a file from Windows and I run one of the several "undelete" utilities (say, TestDisk) I can easily recover the file (because "it's still there but it's marked as deleted"). If I remove the file from Linux (I'm using Ubuntu) no utility can recover the file (unless I use a deep-search signature-based one). Why? It looks like Linux does not just "mark it as deleted" but it wipes away some on-disk structure, is this the case?

    Read the article

  • setreg.exe for Windows 7

    - by victoriah
    I want to use setreg.exe (http://msdn.microsoft.com/en-us/library/aa387700(VS.85).aspx) to disable the certificate revocation check. However, it's based on an older version of .NET than what I have. Microsoft says it's shipped with the older .net SDK, and when I download that, try to install, it says something like 'can't install SDK without .net 1.1'. In the article linked it says for newer versions I should use SignTool, but SignTool does not appear to have the function I need . Is it possible to either - a) find a tool that can perform the function I need that doesn't need me to install the older .NET b) get setreg.exe without downloading the SDK or do I need to install the older .NET on my machine? And if the latter is the case, what do I need to do to install an older .NET? Will it overwrite my current .NET? Thanks,

    Read the article

  • Move entire OS from NTFS drive to bigger ext4 drive.

    - by pangel
    According to SMART data, the hard drive I curently use is about to fail. I bought a new, bigger drive to copy the system to a safer place. The old drive is 160GB. Ubuntu was installed with Wubi, and the partition is NTFS. There are a few other partitions around (recovery partition, swap...) that I don't care about. The new drive is 320GB. I would like the new system to run on ext4, not on NTFS. I looked at solutions that use dd, or clonezilla, but it seems that moving to a different filesystem prevents me from using them. I considered installing a brand new ubuntu on the new hard drive and then copy /home from the old drive to the new drive, but I heard that there would be file permission problems. I would also have to reinstall all my software. One last thing: the NTFS drive has dead sectors. I don't know how this can influence the copy process, but I mention it just in case. edit: I do not care about the windows partition. I just want Ubuntu to make the transition.

    Read the article

  • Network(ing) to the Limit

    - by Oracle OpenWorld Blog Team
     By Karen Shamban While Oracle OpenWorld attendees are networking, there's an Oracle Global IT team that builds and maintains the massive networks that help run the show. The objective? To keep things running as seamlessly and smoothly as possible, constantly evaluate priorities, mitigate risk, and be ready for whatever might happen -- because things do happen when there are 50,000 plus attendees, tens of thousands of devices, unexpected requirements, and a constant flow of up-to-the-minute information. Here's just some of what it takes to keep the conference going, network style: 100 Oracle network, voice, and desktop engineers; security, risk management, and other IT experts, who come in from 17 countries  1000+ network switches 300+ miles of copper and fiber 485 wireless access points 2,500 wired laptops 300 VoIP phones And just where are all these networks and devices deployed? This is what the team had to build and manage: Moscone North, South, and West, including: The keynote hall Oracle DEMOgrounds in the Exhibition Halls Hundreds of session rooms Connection Centers, Social Avenue, Lounges Registration The Howard Street Tent and Taylor Street Cafe tented venues Oracle Square (Union Square) Yerba Buena Gardens Masonic Auditorium Sessions and demos at 8 hotel venues That's a whole lot of networking going on. And here's the kicker: the team has only 4 days to bring get it all up and running across these many venues, and exactly 12 hours to take it all down once the show ends. The Global IT team puts in the equivalent of 152 24-hour days for set-up, 227 24-hour days of support during the conferences, and then tears it all down in about 20 24-hour days. And in case you were wondering, the planning for next year's Oracle OpenWorld starts ... next week. No rest for the weary.  Now THAT's networking!  So hats off to the Global IT team -- the job ain't easy, but somebody's got to do it, and they do it remarkably well.

    Read the article

  • How to lookup a value in a table with multiple criteria

    - by php-b-grader
    I have a data sheet with multiple values in multiple columns. I have a qty and a current price which when multiplied out gives me the current revenue (CurRev). I want to use this lookup table to give me the new revenue (NewRev) from the new price but can't figure out how to do multiple ifs in a lookup. What I want is to build a new column that checks the "Product", "Tier" and "Location/State" and gives me the new price from the lookup table (above) and then multiply that by the qty. e.g. Data > Product, Tier, Location, Qty, CurRev, NewRev > Product1, Tier1, VIC, 2, $1000.00, $6000 (2 x $3000) > Product2, Tier3, NSW, 1, $100.00, $200 (1 x $200) > Product1, Tier3, SA, 5, $250.00, $750 (5 x $150) > Product3, Tier1, ACT, 5, $100.00, $500(5 x $100) > Product2, Tier3, QLD, 2, $150.00, $240 (2 x $240) Worst case, if I just get the new rate I can create another column

    Read the article

  • Redirect with iptables if destination port is not listened

    - by PoltoS
    I've a server listening on port 10000. But this server is running only in a special case (then some third service is available). Otherwise the port is not listened. Is it possible to redirect the client to another port if 10000 is not listening? I see two solutions: 1) insert/remove iptables rules on server start/stop, but since the server may be killed, it may not insert the correct iptable redirect rule before dieing. 2) make a permanent userspace rule that checks if the port is listening and redirects the packet if not listened. How to do 2) ? Do someone have recipes for ipq? May be someone can suggest me a better way? It is something like fallback redirect: I'll have thousand of clients with different ports (10000-11000) and if their instance of server is not running, whey should be redirected to some page explaining why they don't have and instance connected.

    Read the article

  • How to avoid being fooled

    - by Dacav
    I'm a big fan of the OpenSource development model, and I think that sharing information, knowledge and ideas is the best way of working in software development. Still I think that being hired for proprietary software development must not be demonized. Of course, as there's a no reward in terms of sharing there must be a bigger reward in terms of money (i.e. I surrender all my rights for cash). It may happen that one gets hired piecework, for a single project: in this case one is more vulnerable to dishonest employers. This didn't happen to me personally, but some friends of mine had bad experiences, and lost a lot of time without being fairly retributed. Of course a contract should protect both parts. But contracts can be very generic in the specification. Software is not anything but a palpable good! Besides I don't think that contracts can distinguish between a well written software and a poorly written one. Note also that, in this (nasty) spirit, it's also likely that the employer cannot trust the employee! So also the employer should be protected by a dishonest employee). My question is the following: Which is, in your opinion, a good way of avoiding this kind of situation from the technical perspective?

    Read the article

  • Grub2 menu after crash

    - by poopa
    Hi, I have ubuntu 9.10 desktop VMware VM with the default grub2 installed. There is some weird problem with this VM. When you clone this vm and have a customization script run, the cloned machine crashes at first boot (VMware does not officially suport customizaing Ubuntu newer than 8.04). After the creash the Grub boot menu is displayed but there is not time out. I checked /boot/grub/grub.cfg and it does indeed show a timeout of 10 seconds. Nothing happens till I select an option with the keyboard. The second time the Ubuntu loads, it does not crash. My question is, how do I make the grub menu timeout in that case? Thanks.

    Read the article

  • Ubuntu 12.10 and nVidia drivers don't like each other?

    - by mingos
    I decided to upgrade both my computers from Precise to Quantal. What a mistake that was. My laptop has a nVidia GT 330M card, while the desktop has an nVidia 9600 GT. In both cases everything goes great as long as I use the Nouveau driver (ugh!). Can't really play games (Amnesia... and hoping for Steam Beta participation...), even though it's OK for work. Now, ever since 9.04 or so, I just installed nvidia-current and all just worked. Since 12.10, after installing nVidia drivers, Unity won't start at all (hangs with only the wallpaper displayed, no cursor or widgets), Gnome Shell is permanently in fallback mode. Now, I have tried on both computers, with multiple clean installs on Ubuntu (two separate downloads, just in case), one from Ubuntu Gnome Remix. And additionally, Fedora 17, which seems to suffer from the same issue. Tried all nVidia driver suggestions available in Software Sources, and even compiled the drivers myself. I tried several versions of the driver to exclude an issue with the newest one. In my frustration, I have switched to Windows (which, ironically, "just works" with my hardware), but still hold a twin OS configuration on the desktop and would like to use Ubuntu again. So, can anyone point me to where the issue might lie?

    Read the article

  • Release Notes for 5/31/2012

    Here are the notes for this week’s release: Pull Requests We’ve moved the pull request creation UI from a modal dialog to a full page to make room for more context on what you’re sending with your pull request. In this week’s changes, you’ll see the list of commits; look for diffs to appear in the coming weeks. Corrected several issues where the line that had comments associated was not being properly differentiated in the code snippet previews on pull request discussions. Corrected an issue where pull request line comments would not appear to be associated with the proper line of code. Fixed an issue where accepting pull requests was not working properly. Bug Fixes Fixed an issue where code snippets could not be inserted using Opera. Fixed an issue where Internet Explorer displayed an error message when browsing the site using https. Fixed an issue where the CAPTCHA was not displaying properly on the Contact User page. Fixed an issue where paging was not working properly in the “Recent Release in Followed Projects” sections for logged in users. Other Updated the activity stream to no longer auto-load new items when scrolling down. Added an explicit download link at the end of the download countdown in case certain browsers/scenarios don’t automatically start downloading. Have ideas on how to improve CodePlex? Visit our ideas page! Vote for your favorite ideas or submit a new one. Got Twitter? Follow us and keep apprised of the latest releases and service status at @codeplex.

    Read the article

  • Recommended design pattern for object with optional and modifiable attributtes? [on hold]

    - by Ikuzen
    I've been using the Builder pattern to create objects with a large number of attributes, where most of them are optional. But up until now, I've defined them as final, as recommended by Joshua Block and other authors, and haven't needed to change their values. I am wondering what should I do though if I need a class with a substantial number of optional but non-final (mutable) attributes? My Builder pattern code looks like this: public class Example { //All possible parameters (optional or not) private final int param1; private final int param2; //Builder class public static class Builder { private final int param1; //Required parameters private int param2 = 0; //Optional parameters - initialized to default //Builder constructor public Builder (int param1) { this.param1 = param1; } //Setter-like methods for optional parameters public Builder param2(int value) { param2 = value; return this; } //build() method public Example build() { return new Example(this); } } //Private constructor private Example(Builder builder) { param1 = builder.param1; param2 = builder.param2; } } Can I just remove the final keyword from the declaration to be able to access the attributes externally (through normal setters, for example)? Or is there a creational pattern that allows optional but non-final attributes that would be better suited in this case?

    Read the article

  • Ubuntu server 9.10 freezes up after ~10 minutes - unconventional

    - by L33tCh
    There is a topic by essentially the same name, but my answer leads to another question which I would like answered so... My case is as follows: I connect through SSH to my server... no problems for many months on Ubuntu Server 9.04. Eventually I decided to updgrade and straight away I get the 10 minutes freeze problem. I went through all the possible fixes (many can be found here but my quick fix, the only that worked, leaves me with no real answer as to why and how to properly correct it... As ridiculous as it sounds... while I leave a keyboard plugged in... it runs without flaw. Can anyone make sense of this? For the sake of completeness, (although none of the following made any difference upon subsequent completion), some of the permanent changes I have made while trying to correct the issue are: replacing samba config file with 9.04 backup, booting with noapic and modifying sysctl.conf, just uncommenting additional settings for enhanced security.

    Read the article

  • process ksoftirqd consumes permanent 15% CPU load [closed]

    - by markus
    Possible Duplicate: Anyone else experiencing high rates of Linux server crashes during a leap second day? The process ksoftirqd/0 uses permanent 15% CPU on our debian squeeze server. 4 root 20 0 0 0 0 R 15.0 0.0 850:59.17 ksoftirqd/0 I already read that this can have various reason like Full harddisk or high network traffic. In our case we do have more or less low network traffic and enough space on hard disk. How can I analyse what causes ksoftirqd/0 to use permanently 15% CPU?

    Read the article

  • Configure a app in a subpath in nginx

    - by Rubem Azenha
    I have a nginx + passenger + single rails app setup in a Fedora server. My nginx.conf regarding my app looks something like this: server { listen 80; server_name myapp.mycompany.com; root /opt/apps/myapp_large_name/public; passenger_enabled on; client_max_body_size 4M; } So when I access myapp.mycompany.com.br it access to my rails app under /opt/apps/myapp_large_name/public. I want to be able to configure a subpath to access to another app (in this case, monit). I want something like that: myapp.mycompany.com.br/monit access the monit app. Is it possible?

    Read the article

< Previous Page | 462 463 464 465 466 467 468 469 470 471 472 473  | Next Page >