Search Results

Search found 16537 results on 662 pages for 'oracle accelerate for midsize companies'.

Page 598/662 | < Previous Page | 594 595 596 597 598 599 600 601 602 603 604 605  | Next Page >

  • How can the shared hosting server provide unlimited physical subdomains as opposed to unlimited virtual subdomains?

    - by xport
    Some hosting companies offer unlimited subdomains. There are two kind of subdomains: physical subdomains and virtual subdomains. A physical subdomains has its own site directory rather than being nested inside the site directory of its parent domain. A virtual subdomain site directory, on the other hand, is nested inside the site directory of its parent domain. I wonder how can the shared hosting company provide unlimited (theoritically) physical subdomains? In my understanding, each physical subdomain represents a new site (rather than a new application or virtual directory) in IIS. Please correct me if my mental model is wrong.

    Read the article

  • Computer Loses connection to router

    - by ajon
    I have a new Windows 8 Dell XPS desktop computer at work. It has a Dell Wireless 1703 802.11 b/g/n (2.4ghz) wireless adapter. sporadically throughout the day I lose internet connectivity. When I click on the wireless system tray icon, my companies router does not appear in the list. No other computer on the network is losing connection, only this one. I don't even know what to check. Could it be an ip conflict that causes me to get booted off? (I use DHCP), but why wouldn't my computer be able to see the network. The next time it happens I will see what happens when I try to access the router directly.

    Read the article

  • ISA bus on newer computers

    - by Kevin Ivarsen
    Are there companies that sell new computers that support old ISA bus expansion cards? We have an aging computer running DOS that operates some machinery via an ISA interface board. Updated versions of this board (e.g. PCI, USB) are not available, and I am concerned about the long-term reliability of the 8+ year old computers we currently keep around as backups. If these newer ISA-capable machines exist, are there any general gotchas to be aware of in terms of compatibility with older expansion boards, ability to run DOS, etc.?

    Read the article

  • What kind of proxy acl rules should be applied?

    - by user42891
    I try to block sites in squid based on this article. Assuming you would want to block access to Yahoo (e.g http://www.yahoo.co.jp, http://www.yahoo.com, http://www.yahoo.co.in), you would ideally want to block all of the above URLs, if I use a regular expression and try to search something called yahoo it seems to get blocked. We are just interested in applying rules which would be most commonly used across all companies (e.g. social networking sites like facebook, orkut), porn sites (e.g. sex), gaming sites (games), movie & song download sites, and sites where they can upload data (e.g. rapidshare) What would be the common set of effective rules in achieving the above?

    Read the article

  • How to get a "live screenshot" in windows

    - by alexei7
    Quick question – do you know any software or way of doing a live screenshot of a window and bring it to the foreground? It’d be like picture link in Excel... My problem is that I am running various tasks in IE and Oracle, and they take forever but i have no way of seeing if they’re done without tabbing back into them all the time. I want something so that I can always see in the foreground the status bar/egg-timers thingy for them. Thanks all

    Read the article

  • Which version of ZFS allows shrinking of a pool?

    - by George Bailey
    I found a list of versions and their Solaris release numbers http://download.oracle.com/docs/cd/E19253-01/819-5461/appendixa-1/index.html I know that you can grow a pool by replacing drives with larger ones or adding new drives or mirrors to the pool. I heard that ZFS did not yet support shrinking pools by removing drives/mirrors. But that has probably been changed. Which version (if any) released the ability to shrink a pool?

    Read the article

  • options for deploying application

    - by terence
    I've created a simple web application, a self-contained tool with a user system. I host it publicly for everyone to use, but I've gotten some requests to allow companies to host the entire application privately on their internal systems. I have no idea what I'm doing - I have no experience with deployment or server stuff. I'm just some person who learned enough JS and PHP to make a tool for my own needs. The application runs with Apache, MySQL, and PHP. What's the best way to package my application to let others run it privately? I'm assuming there's better options than just sending them all the source code. I'd like to find a solution that is: Does not require support to set up (I'm just a single developer without much free time) Easy to configure Easy to update Does there exist some one-size-fits all thing that I can give to someone, they can install it, and bam, now when they go to http://myapplication/ on their intranet, it works? Thanks for your help.

    Read the article

  • Can you set up a Linode VPS as a reseller type scenario?

    - by MAZUMA
    First, I'm a fairly green on this subject. So, go easy. I do web design and development and would like to set up a way to be the central host so to speak for my clients. Currently they are scattered across multiple hosting companies. I've looked at some the cheapy options like a LunarPages and some of the more robust options like RackSpace Cloud and Linode. RackSpace has a reseller program, but it's managed and more at the top end on cost. I'm not seeing any type of reseller program in place for Linode. Can I purchase a higher Linode plan and use Apache2's VirtualHost configuration to handle multiple domains (about 8 domains, but hopefully will grow.). Or, am I not thinking about this correctly?

    Read the article

  • Distributed website server redundancy

    - by Keith Lion
    Assume a website infrastructure is very complicated and is fully distributed (probably like most large web companies). Am I right in thinking that although there are all these extra web servers to handle multiple client requests, there is still a single "machine" whereby users must enter? I am guessing this machine will be the one physically associated to the IP address? I ask because I need to know whether, in places where distributed systems exist, there is still a single point of failure- usually the control node or, in this example, the machine connected to the public internet? Surely there cannot be two machines connected to the internet, as they would have to have different IP addresses? This "machine" may not be a server per se, but maybe it is a piece of cisco equipment. I just need to know whether, in the real world, these distributed systems still have a particular section where they depend on the integrity of one electronic device?

    Read the article

  • How to whitelist a domain while blocking forgeries using that domain?

    - by QuantumMechanic
    How do you deal with the case of: wanting to whitelist a domain so that emails from it won't get eaten, but not having emails forged to appear to be from that domain get bogusly whitelisted whitelist_from_recvd looks promising, but then you have to know at least the TLD of every host that could send you mail from that domain. Often RandomBigCompany.com will outsource email to one or more sending companies (like Constant Contact and the like) in addition to using servers that reverse-resolve to something in its own domain. But it looks like whitelist_from_recvd can only map to one sending server pattern so that would be problematic. Is there a way to say something like "if email is from domain X, subtract N points from the spam score"? The idea would be that if the mail is legit, that -N will all but guarantee it isn't considered spam. But if it is spam, hopefully all the other failed tests will render it spam even with the -N being included.

    Read the article

  • How to specify root's environment variable?

    - by Wendy
    I do rails development. In this app, I need to specify the environment variable LD_LIBRARY_PATH = /usr/local/oracle/lib But when I run the app with sudo script/server, it doesn't run because that library path is not in roots' env. What should I do to make it work? I tried to put the path under root ./bashrc and it didn't work.

    Read the article

  • Dealing with employees that use Hotspots? [closed]

    - by javasocute
    Recently a consultant revealed to me and my superiors that a particular department in my company had a tremendous decline in productivity over the past year. This department happened to mostly work with laptops and tablets and thus had wireless. Our barracuda did not display any vast amounts of internet entertainment and so I was initially puzzled by his statement. However he explained that more and more employees are using "hotspots" that when not broadcasted, become invisible to our network. So basically these people could connect to hotspots and watch youtube all day and we would never know it. The consultant also said that this was becoming an issue for many companies. So my question is, how can I detect invisible hotspots? I was browsing google and there are several different programs that can scan for such access points. But I am hoping to get a recommendation from the serverfault community. Thank you.

    Read the article

  • MySQL for SQL Server DBAs

    - by SQL3D
    I've been tasked with taking over the administration of a MySQL installation (running on Red Hat Linux) that will become fairly critical to our business in the near future. I was wondering if anyone could recommend some resources in regards to administering MySQL for DBAs already experienced with other relational database (SQL Server and some Oracle in my case). Specifically I'm looking for information around disaster recovery as well as high availability to start with, but I do want to get well rounded with the entire system. Thanks in advance, Dan

    Read the article

  • Directing multiple domains to one server

    - by dtechie
    Hi I would like to host 5 word-press blogs on my unlimited server space i bought from a web-hosting company similar to hostgator. Now the companies say it is technically not possible to direct more than one domain to the hosting package and i would need to buy multiple hosting packages . I read about vhost (apache),domain alias,ip forwarding But not sure how ask my webhost about it when they say it is not possible here is what i want to do xhost.com/folder1 << www.x.com xhost.com/folder2 << www.y.com xhost.com/folder3 << www.z.com xhost.com/folder4 << www.aa.com the web host has a cpanel so dont know if they give/have access to vhosts if they are a reseller thank you for your help

    Read the article

  • First-class help desk solution? [closed]

    - by Andy Gregory
    A help desk, for larger companies is a place for centralized help within an enterprise to help the users of their products and services.Therefore,to find the solution that best fits your business requirements, it is important to research, examine, and compare help desk software. As far as I know,Hesk is an free helpdesk software with some limited features while h2desk can provide the hosted solution. But for my small business,i just need a web based help desk software which can provide ticket management and knowledge base faq. We need unlimited staff support. Maybe a freeware help desk software can not meet our needs. So,we are willing to pay for effective helpdesk solution. But it should be low cost. We have gotten two choice: iKode Helpdesk Nethelpdesk Our helpdesk team are tend to iKode Helpdesk. Any other efficient first-class help desk solution to share?

    Read the article

  • What is the world wide web? [closed]

    - by think123
    I don't know where to post this question, so please move it if necessary. Ok, so I've heard of how the professional hosting companies can create 'links' to the world wide web to register an unregistered domain. So that's where my question comes from. Is the world wide web a server to which servers link? Is it created by abstract linkage? I'm not sure. Also, what does it mean for the DNS to be updated throughout the whole world?

    Read the article

  • How to make a dropdown list such that... (see details)

    - by daysandtimes
    I want to plot the stock prices of certain companies VS the S&P500. I have all the price data downloaded in my excel sheet already. I want to create a line graph in Excel. One line is fixed and that would be the S&P500, and the other would be the company I select. I know how to use data validation to create a dropdown list, but how could I make it in such a way that when I select company A, I would only see company A's price data but not company B, C, etc. Then when I select company B, I would only see company B's price data & the S&P price data. And so on. The S&P line will be visible all the time, but the various company's price line will only appear if it is selected. Is there any easy way to normalize all sets of data in a way such that the starting point is always 100?

    Read the article

  • Why can't we reach some (but not all) external web service via VPN connection?

    - by Paul Haldane
    At work (UK university) we use a set of Windows servers running WS2008R2 and RRAS which offer VPN service to students in our accommodation. We do this to associate the network connections with individuals. Before they've connected to the VPN all they can talk to is the stuff thats needed to setup the VPN and a local web site with documentation on how to connect. Medium term we'll probably replace this but it's what we're using at the moment. VPN on the 2008 servers allocates client a private (10.x) address. Access to external sites is through NAT on the campus routers (same as any other directly connected client on a private address). Non-VPN connections aren't seeing this problem. Older servers run WS 2003 and ISA2004. That setup works but has become unreliable under load. Big difference there was that we were allocating non-RFC1918 addresses to the clients (so no NAT required). Behaviour we're seeing is that once connected to the VPN, clients can reach local web sites (that is sites on the campus network) but only some external sites. It seems (but this may be chance) that the sites we can reach are Google ones (including YouTube). We certainly have trouble reaching Microsoft's Office 365 service (which is a pain because that's where mail for most of our students is). One odd bit of behaviour is that clients can fetch (using wget on a Windows 7 client) http://www.oracle.com/ (which gets a 301 redirect) but hangs when asked to fetch http://www.oracle.com/index.html (which is what the first URL redirects to). Access works reliably if we configure clients to use our local web proxies (Squid). My gut tells me that this is likely to be something in the chain dropping replies either based on HTTP inspection or the IP address in the reply. However I'm puzzled about why we're seeing this with the VPN clients. Plan for tomorrow (when I'm back in the office) is to setup a web server on external connection so that we can monitor behaviour at both ends of the conversation (hoping that the problem manifests itself with our test server). Any suggestions for things we should be looking at?

    Read the article

  • Losing connection to server from several computers

    - by user3696358
    I have a PostgreSQL server running Oracle Linux. In my network I have several workstations each running different OS (Fedora, Ubuntu, Windows). Every once in a while I lose connection from the several workstations to the server while other workstations can connect with no problem. If I do service network restart from the server the problem is solved and everyone can work until the next time it returns. Any clues? Thank you, Ben.

    Read the article

  • When RAID 10 is SLOWER than RAID 1, why?

    - by Paul
    We have a Dell 2950 with PERC and 14 external SAS 15K 73GB drives. An Oracle database job takes 3 hours to run with the drives set as hardware RAID 10 (striped across 7 mirrored pairs). The same job with the drives in RAID 1 takes only 1 hour. OS is Win 2008 R2 I think. Before we change the RAID level (with considerable downtime) on the production box, does anyone know why we're seeing this odd result, and if there's a better way to fix it?

    Read the article

  • DKIM sign any outgoing mail as the same domain. is possibly?

    - by Felipe Alcacibar
    I have a website, and the users recommend via email to another persons, so the mail come from my site, but I need to appear the user as sender, but "via" mydomain. The logically and what lists management software and companies do is differ the "MAIL FROM" SMTP command and the "From" header (like mailchimp, for example). So I want to DKIM sign a mail with mydomain.com, even if the From header says another domain. is possibly this? I am using postfix with Debian GNU/Linux, and in this case using dkim-filter and dk-filter.

    Read the article

  • Forward name server for faster load times.

    - by user54017
    I have a website hosted in China, which must be hosted there for various reasons. I also cannot change hosting companies for various reasons. It is a website that is accessed worldwide, however it has atrocious ping rates from everywhere but China. This is just an idea, so this may be impossible and/or not make any sense, but here it goes. A friend of mine has used proxies which helped his ping and load times from servers in Korea. Is there any way I can do this with my server? Perhaps a service which forwards my name servers to somewhere with much better global ping rates? Any other ideas which will make pings/load times fester other than this are greatly appreciated as well. Please keep in mind when answering: I cannot change servers I do not have access to configure the server itself.

    Read the article

  • Where to run Java and PHP code continuously

    - by az1112
    I'm sorry if this is the wrong forum for a question like this, but you have to start asking questions somewhere to get anywhere :) The question is pretty simple: I need a server where I can run Java and PHP scripts 24 hours a day, 7 days a week. I need to be able to access this server via SSH, and I need to be able to retrieve the data generated by scripts using SCP. Also, I need to be able to run 10-20 scripts simultaneously. What is the name of the thing I should be looking for? Is it a dedicated server? I'm confused for 2 reasons: 1) because there seem to be all kinds of servers out there; 2) because most companies advertising dedicated servers seem to be aiming them at people who want to host websites. But I don't want to host a website; I just want to run my code.

    Read the article

  • Is it possible to create a full "encrypted worried about privacy" VPS, but still being externally us

    - by Somebody still uses you MS-DOS
    I've been reading a lot of things about privacy, "being in control of your data" and everything, and now a project called diaspora* is trying to be an OSS Facebook alternative. Fact is: you still have to use a server. Even if you use a VPS somewhere, they still have access do your data, so diaspora* isn't that protective shell people are looking for absolute power over your data unless you create a server on your basement. My question is: is it possible to create a really encrypted usable server using a VPS? From database to source files? If not, what it can be "obfuscated" or encrypted? (And just a mention, not really my question, do you think is diaspora* really possible to be made?) (I know if you really want privacy you shouldn't be even using these services and being social, but I'm asking if it's possible to at least avoid companies using your data)

    Read the article

  • How long does it take in practice to warm up large in-memory databases?

    - by Sim
    Companies such as Peak Hosting are offering 64 core machines with 512Gb RAM for $2K/month. This is a very interesting choice for in-memory databases such as Memcached/Redis as well as databases whose performance degrades rapidly when the data & indexes don't fit in RAM, such as MongoDB. My main concern with monster machines such as these is the time it takes to warm up an in-memory database. In my experience, theoretical metrics, e.g., that SATA can load 100Mb/sec, fall short of what happens in practice. Even at that rate, 100Mb/sec means that loading up 512Gb RAM machine from SATA disks can take over 1 1/2 hours (!). I am looking for real-world reports of warm-up times for machines with very large memory. Please, share details of the software on the machine, data size, storage configuration, e.g., SATA or SSD, network, hosting/cloud provider, if relevant, etc.

    Read the article

< Previous Page | 594 595 596 597 598 599 600 601 602 603 604 605  | Next Page >