Search Results

Search found 59995 results on 2400 pages for 'web experience management'.

Page 403/2400 | < Previous Page | 399 400 401 402 403 404 405 406 407 408 409 410  | Next Page >

  • Teredo and IPV6 web servers

    - by Sandro Antonucci
    Hello, I have two different questions, the first one could be stupid. I was reading about IPV6 and ended up in this site http://test-ipv6.com/ that says that I have a public IPV6 with Teredo tunneling. I don't understand if this Teredo is a software thing, does it have to with my ISP, is it enabled by Windows, how do I get that IPV6 IP, where does it come from? I have windows 7. Second question, I have a VPS with IPV6 connectivity only I can connect to it (apparently thanks to Teredo) and visit pages from the web server using IPV6 IP directly into the browser, is it possibile to host a website on there accessible by the IPV4 network? maybe some dns servers than tunnel traffic from IPV4(user) to IPV6(server)? Thank you

    Read the article

  • Apache security for multi-user development web server.

    - by mrmartinblue
    I've been searching and reading through documents all morning and understand that I need to use some combination of chown and probably 'jailing' to securely give programmers access to directories on my centos webserver. Here's the situation: I have an apache web server that has any number of virtual sites located in /var/www/site1 /var/www/site2 etc.. I have different developers that need full access both ssh and vsFTP to only the site they are working on. What is the best way to create and maintain security in this scenario. My thought would be to create a new user for each coder, jail that user to the website directory they are allowed to work in, add their user to a group and set the webroot's owner to that group. Any thoughts? Good, bad, ugly? Thanks!

    Read the article

  • Software for formatting / web-publishing tutorial-type, technical articles

    - by gojira
    I want to put some of my own technical writing on the web: tutorials, how-to's, conceptual articles about programming etc. I write mostly for myself yet, but think others can benefit from it so I want to put it online, and I also have plans of making everything more coherent and turning it into a book. My articles are not meant as a blog, they're just meant as a collection of tutorials and how-to's, but I think a blog-type software would probably be the most appropriate mode of formatting that kind of stuff. Right now I am using the blog feature of cite u like, which is very low tech and I really need something better. On my wish list is, among other things: code snippets in scrollable boxes syntax highlighting for source code snippets tagging articles with graphical icons according to topic, like in slashdot of course users (i.e. readers) should be able to comment on articles Also, just more options than a super-simple blog should look slick! should not look wikipedia-like Is there any software you can recommend for this purpose?

    Read the article

  • What's the best platform to publish documentation to internal users?

    - by serialhobbyist
    My team has a need to publish documentation internally. At the moment, it's spread all over the place and this means we often have to search everywhere to find something. We'd like to publish everything in one place. The main thing that stops us is access control - the wikis in place don't belong to us and we can't do it. What is the best tool for publishing docs, ideally fitting these requirements: web front end - readers access docs using browser single place to put docs access control by individual doc or by sets of docs (folders, branch of 'site', ...) if you don't have access to a doc, you don't see the link to that page/doc/folder. either built-in editor or something my users are familiar with (e.g. Word) built-in version control would be nice Also, can you think of other criteria I should've specified?

    Read the article

  • how to connect public web server to internal LAN

    - by DefSol
    I have a VPS which is my public web server for all my clients. It's running server 2008 and I would like to have it connect via secure connection to my internal LAN. I would like this to be a route so access is bi-derectional. Have read about Server & Domain isolation, but am concerned this may prevent public views to the webs sites on the server. I currently have a PPTP tunnel, but I'm wanting better security (IPSec or SSL etc) and it's not given my bi derectional access. (In fact my backups aren't copying accross but this could be an acl issue) The goal is to provide easy/automated backups of data & sql db's to my internal LAN, as well as a means to provision new sites & db's from a workflow occuring internally. Internal lan is windows based with ISA 2006 at the perimeter. Thanks

    Read the article

  • Generating/managing config files for hosted application

    - by mfinni
    I asked a question about config management, and haven't seen a reply. It's possible my question was too vague, so let's get down to brass tacks. Here's the process we follow when onboarding a new customer instance into our hosted application : how would you manage this? I'm leaning towards a Perl script to populate templates to generate shell scripts, config files, XML config files, etc. Looking briefly at CFengine and Chef, it seems like they're not going to reduce the amount of work, because I'd still have to manually specify all of the changes/edits within the tool. Doesn't seem to be much of a gain over touching the config files directly. We add a stanza to the main config file for the core (3rd-party) application. This stanza has values that defines the instance (customer) name the TCP listener port for this instance (not one currently used) the DB2 database name (serial numeric identifier, already exists, they get prestaged for us by the DBAs) three sub-config files, by name - they need to be created from 3 templates and be named after the instance The sub-config files define: The filepath for the DB2 volumes The filepath for the storage of objects The filepath for just one of the DB2 volumes (yes, redundant to the first item. We run some application commands, start the instance We do some LDAP thingies (make an OU for the instance, etc.) We add a stanza to the config file for our security listener that acts as a passthrough to LDAP instance name LDAP OU TCP port for instance DB2 database name We restart the security listener (off-hours), change the main config file from item 1, stop and restart the instance. It is now authenticating via LDAP. We add the stop and start commands for this instance to the HA failover scripts. We import an XML config file into the instance that defines things for the actual application for the customer - user names, groups, permissions, and business rules. The XML is supplied by the implementation team. Now, we configure the dataloading application We add a stanza to the existing top-level config file that points to a new customer-level config file. The new customer-level config file includes: the instance (customer) name the DB2 database name arbitrary number of sub-config files, by name Each of the sub-config files defines: filepaths to the directories for ingestion, feedback, backup, and failure those filepaths have a common path to a customer-specific folder, and then one folder for each sub-config file Each of those filepaths needs to be created We need to add this customer instance to our monitoring scripts that confirm the proper processes are running and can be logged into. Of course, those monitoring config files include the instance name, the TCP port, the DB2 database name, etc. There's also a reporting application that needs to be configured for the new instance. You get the idea. There's also XML that is loaded into WAS by the middleware team. We give them the values for them to plug into the XML - they could very easily hand us the template and we could give them back completed XML.

    Read the article

  • Are there any HTPC-optimized web browsers?

    - by smackfu
    Features that I would ideally include in "HTPC-optimized": Full-screen. Navigable using a remote or the keyboard arrows. Legible at couch distances. Or, to put it another way, imagine the design requirements for Hulu Desktop or XBMC or WMC, applied to a web browser. Opera on a Wii meets most of these criteria but not being HD wastes a lot of potential. If a single solution doesn't exist, is there some combination of Firefox add-ons that will get me there?

    Read the article

  • Web connection speed issue

    - by Iroll750
    I encounter a strange issue with my web (wireless) connection on my laptop (Ubuntu 11.10) I'm able to have 540kbps (down) on all computers in my house (including laptop at the same distance from the box) except me. When i do a speed connection test, i can have this optimal speed immediately after booting. As i surf on the net, watching videos, download things, etc... it becomes slower progressively until 100-110 kbps and the speed remain the same until i reboot. I tried to figure out if a process consume bandwith in background with nethogs but i found nothing suspicious. Any ideas ? Thanks in advance for your help !!

    Read the article

  • Backing up data stored on Amazon S3

    - by Fiver
    I have an EC2 instance running a web server that stores users' uploaded files to S3. The files are written once and never change, but are retrieved occasionally by the users. We will likely accumulate somewhere around 200-500GB of data per year. We would like to ensure this data is safe, particularly from accidental deletions and would like to be able to restore files that were deleted regardless of the reason. I have read about the versioning feature for S3 buckets, but I cannot seem to find if recovery is possible for files with no modification history. See the AWS docs here on versioning: http://docs.aws.amazon.com/AmazonS3/latest/dev/ObjectVersioning.html In those examples, they don't show the scenario where data is uploaded, but never modified, and then deleted. Are files deleted in this scenario recoverable? Then, we thought we may just backup the S3 files to Glacier using object lifecycle management: http://docs.aws.amazon.com/AmazonS3/latest/dev/object-lifecycle-mgmt.html But, it seems this will not work for us, as the file object is not copied to Glacier but moved to Glacier (more accurately it seems it is an object attribute that is changed, but anyway...). So it seems there is no direct way to backup S3 data, and transferring the data from S3 to local servers may be time-consuming and may incur significant transfer costs over time. Finally, we thought we would create a new bucket every month to serve as a monthly full backup, and copy the original bucket's data to the new one on Day 1. Then using something like duplicity (http://duplicity.nongnu.org/) we would synchronize the backup bucket every night. At the end of the month we would put the backup bucket's contents in Glacier storage, and create a new backup bucket using a new, current copy of the original bucket...and repeat this process. This seems like it would work and minimize the storage / transfer costs, but I'm not sure if duplicity allows bucket-to-bucket transfers directly without bringing data down to the controlling client first. So, I guess there are a couple questions here. First, does S3 versioning allow recovery of files that were never modified? Is there some way to "copy" files from S3 to Glacier that I have missed? Can duplicity or any other tool transfer files between S3 buckets directly to avoid transfer costs? Finally, am I way off the mark in my approach to backing up S3 data? Thanks in advance for any insight you could provide!

    Read the article

  • As a student looking for hands-on experience, how should I configure my test lab to accurately reflect a business setting?

    - by Moe
    I'm one class away from my BA IT, I took several classes in general IT. Out of all the books I found just two to be really beneficial, so I'm trying to get the hands on experience. I want to build a small test network that has both wireless and also wired clients, a printer, a laptop, a desktop, and server (I have 2x 1TB drives of data that I want to be available to all computers). What should I do to configure these in a way that will reflect how a small business might be set up, so that I can work towards some meaningful experience.

    Read the article

  • Cant deploy "war" file from Virtual Hosts, see a directory listing.

    - by Kaustubh P
    This is my httpd.conf configured with Virtual hosts: NameVirtualHost *:80 <VirtualHost *:80> ServerName http://foo.baz.in DocumentRoot /var/www/foo/ </VirtualHost> <VirtualHost *:80> ServerName http://bar.baz.in DocumentRoot /var/www/ </VirtualHost> The second virtual host is a Wordpress blog, configured with .htaccess, and index.php in the root i.e. /var/www, and rest of the files in wordpress's own folder. However, the first virtual host is a "war" file, and when I goto foo.baz.in, I see the directory listing, containing the war. I also tried changing the DocumentRoot to /var/www/foo/foo.war` but I get an error Restarting web server: apache2Warning: DocumentRoot [/var/www/foo/foo.war] does not exist I also changed the owner and permission of the war to www-data:www-data and changed the permissions to 755, but to no avail. How do I make apache deploy my "war"? Thanks.

    Read the article

  • How to supply parameters in URI schema?

    - by abhishekgarg
    I just started working with URI schema and successfully created one in Windows and Linux as well, but I am not able to parse any parameters to it. In Linux I am trying to open a file "test.py" in gedit, so for schema part I used these commands: gconftool -2 -t string /desktop/gnome/url-handlers/geditapp/command "gedit %s" gconftool -2 -t bool/desktop/gnome/url-handlers/geditapp/enabled true This is creating the URI protocol and I'm able to open the application with the Web-browser, but its not taking the parameters for the file I want to open, so I'm using the following command: <a href="geditapp:/opt/test/myfile.py">open</a> Which opens the gedit but without the file. Can someone please help me with this?

    Read the article

  • Using Virtual Domains with WordPress and Cherokee Web Server

    - by ikevinjp
    I'm trying to configure WordPress to work with Cherokee Web Server. The multi-site set up in WordPress is yyy.com and uses 'sub-domains'. Thus, I have sites like aaa.yyy.com, bbb.yyy.com - they all work fine so far. However, I have another completely different domain: abcde.com. Within WordPress, the sub-domain set up is abcde.yyy.com. When a public user accesses abcde.com, it cannot properly be redirected to the WordPress sub-domain abcde.yyy.com. How do I set up Cherokee to properly redirect virtual domains to WordPress? That means, the domain facing public would be abcde.com but WordPress would want to see abcde.yyy.com to properly return the correct content for the specific site. (I tried and it always get redirected to something like this: "http://yyy.com/wp-signup.php?new=abcde.com")

    Read the article

  • Bulk Deleting All Messages in a Folder in Microsoft Outlook Web Access

    - by Chris S
    How do you delete all messages in a folder in Outlook, preferrably through Web Access? I left my Outlook account unattended for several days (on vacation) and when I got back I found several folders with over 5k emails, mostly error logging or spam. When I try to open the Outlook client, it just locks up, presumably unable to download that many emails. I can view at most 100 emails at a time, but I can't select all emails to delete or permanently delete them immediately, so manually deleting this many emails is going to take a while. Gmail has a similar feature to select and delete all emails in a folder, and that's free so I figure being a quality non-free product from Microsoft, Outlook should have a similar feature (yes that's sarcasm). I've Googled, but I'm not finding anything. Is this possible?

    Read the article

  • Could not open IPV6 addr on web browser

    - by vito
    I have ipv6 address fe80::21f:a4ff:fe91:2e44%4. This is the address of modem/router. I am unable to open this address in browser to view the web configuration. I am able to telnet to fe80::21f:a4ff:fe91:2e44%4, view the console UI & can do settings. I can also open 192.168.1.1(IPv4 address of modem/router) in browser. I tried [fe80::21f:a4ff:fe91:2e44] in all browsers like IE, Firefox & Chrome. It was not possible.

    Read the article

  • Mac: Script application downloaded from the Internet

    - by Svish
    I downloaded a php framework and has started to make a website using that. Sometimes I need to look at the source of that framework and every time I open a file I haven't opened before I get this message: “somefile.php” is a script application which was downloaded from the Internet. Are you sure you want to open it? That is ok and nice I suppose, but I am getting tired of it. Is there a way I can fix all the files in my web directory so that the os somehow forgets the files are from the Internet or something like that?

    Read the article

  • IIS6 cannot start additional processes when page request is called from other than localhost

    - by awe
    I have a web application dll that is running under iis6. This is starting a number of processes to be able to handle more than one request at the time. This is working perfect in a numbers of installations. In this particular installation, it runs perfectly when initialized by a call in IE on the server using http://localhost/apppath . The problem is when the processes are from another location, i.e. another computer on the same network initialising the call throught http://servername/apppath . In this case, the initial dll running under iis is executing (proved by logging), but it fails to initialize the additional processes. If the additional processes are already initialized by a call from localhost, it also works when called from another machine (in this case, it is just attached to the existing processes).

    Read the article

  • Public DNS Server fails on Windows Amazon EC2

    - by Adroidist
    I have started a new Windows server instance on Amazon EC2. The security group has the following rules: Ports Protocol Source 22 tcp 0.0.0.0/0 80 tcp 0.0.0.0/0 443 tcp 0.0.0.0/0 3389 tcp 0.0.0.0/0 53 udp 0.0.0.0/0 -1 icmp 0.0.0.0/0 I am able to ping the public DNS server of the machine and i can connect to it using Windows Remote Desktop connection. However, when i put in my web browser the public DNS server, it fails to connect. Morever, I used filezilla and putty (and in both I loaded the private key .pem) but i receive connection timed out. I disabled the firewall on both my pc and the instance (which I entered using Remote desktop connection). Can you please tell me what I am missing?

    Read the article

  • make IIS 7.5 cache static content files over diferent pages

    - by Achilles
    On a Windows 2008 R2, using DNS and IIS I've established my development test server; i.e. I'll have a web application that I can browse on http://test.dev I've moved all the static content files like images, js files and css files into another application which is visible on http://cdn.test.dev test.dev, uses cdn.test.dev urls like http://cdn.test.dev/js/jquery.js to load js, css and images. When I first load "~/" of test.dev, all files will load with a response code of 200; when I press F5 in Firefox, all files, except the "~/default.aspx", will load with 304 response code; but pressing Ctrl+F5 loads them again with a 200 code; if I browse another url like "~/pages/" in test.dev, all of those static files will reload with a 200 code... Is this normal or I'm doing something wrong? Actually, I'm looking for a behavior like this: I want the client to load http://cdn.test.dev/js/jquery.js, only once. I want the client's browser to use this jquery.js file, from cache, in all other pages of test.dev Is this possible? This is the web.config file I have in the root directory of cdn.test.dev: <configuration> <system.webServer> <caching> <profiles> <add extension=".png" policy="CacheUntilChange" varyByHeaders="User-Agent" location="Client" /> <add extension=".gif" policy="CacheUntilChange" varyByHeaders="User-Agent" location="Client" /> <add extension=".jpg" policy="CacheUntilChange" varyByHeaders="User-Agent" location="Client" /> <add extension=".js" policy="CacheUntilChange" varyByHeaders="User-Agent" location="Client" /> <add extension=".css" policy="CacheUntilChange" varyByHeaders="User-Agent" location="Client" /> <add extension=".axd" kernelCachePolicy="CacheUntilChange" varyByHeaders="User-Agent" location="Client" /> </profiles> </caching> <httpProtocol allowKeepAlive="true"> <customHeaders> <add name="Cache-Control" value="public, max-age=31536000" /> </customHeaders> </httpProtocol> <validation validateIntegratedModeConfiguration="false" /> <modules runAllManagedModulesForAllRequests="true"> <remove name="RadUploadModule" /> <remove name="RadCompression" /> <add name="RadUploadModule" type="Telerik.Web.UI.RadUploadHttpModule" preCondition="integratedMode" /> <add name="RadCompression" type="Telerik.Web.UI.RadCompression" preCondition="integratedMode" /> </modules> <handlers> <remove name="ChartImage_axd" /> <remove name="Telerik_Web_UI_SpellCheckHandler_axd" /> <remove name="Telerik_Web_UI_DialogHandler_aspx" /> <remove name="Telerik_RadUploadProgressHandler_ashx" /> <remove name="Telerik_Web_UI_WebResource_axd" /> <add name="ChartImage_axd" path="ChartImage.axd" type="Telerik.Web.UI.ChartHttpHandler" verb="*" preCondition="integratedMode" /> <add name="Telerik_Web_UI_SpellCheckHandler_axd" path="Telerik.Web.UI.SpellCheckHandler.axd" type="Telerik.Web.UI.SpellCheckHandler" verb="*" preCondition="integratedMode" /> <add name="Telerik_Web_UI_DialogHandler_aspx" path="Telerik.Web.UI.DialogHandler.aspx" type="Telerik.Web.UI.DialogHandler" verb="*" preCondition="integratedMode" /> <add name="Telerik_RadUploadProgressHandler_ashx" path="Telerik.RadUploadProgressHandler.ashx" type="Telerik.Web.UI.RadUploadProgressHandler" verb="*" preCondition="integratedMode" /> <add name="Telerik_Web_UI_WebResource_axd" path="Telerik.Web.UI.WebResource.axd" type="Telerik.Web.UI.WebResource" verb="*" preCondition="integratedMode" /> </handlers> <security> <requestFiltering> <requestLimits maxAllowedContentLength="10485760" /> </requestFiltering> </security> <staticContent> <clientCache cacheControlMode="UseExpires" httpExpires="Wed, 01 Jan 2020 00:00:00 GMT"/> </staticContent> </system.webServer> <appSettings /> <system.web> <compilation debug="false" targetFramework="4.0" /> <pages> <controls> <add tagPrefix="telerik" namespace="Telerik.Web.UI" assembly="Telerik.Web.UI" /> </controls> </pages> <httpHandlers> <add path="ChartImage.axd" type="Telerik.Web.UI.ChartHttpHandler" verb="*" validate="false" /> <add path="Telerik.Web.UI.SpellCheckHandler.axd" type="Telerik.Web.UI.SpellCheckHandler" verb="*" validate="false" /> <add path="Telerik.Web.UI.DialogHandler.aspx" type="Telerik.Web.UI.DialogHandler" verb="*" validate="false" /> <add path="Telerik.RadUploadProgressHandler.ashx" type="Telerik.Web.UI.RadUploadProgressHandler" verb="*" validate="false" /> <add path="Telerik.Web.UI.WebResource.axd" type="Telerik.Web.UI.WebResource" verb="*" validate="false" /> </httpHandlers> <httpModules> <add name="RadUploadModule" type="Telerik.Web.UI.RadUploadHttpModule" /> <add name="RadCompression" type="Telerik.Web.UI.RadCompression" /> </httpModules> <httpRuntime maxRequestLength="10240" /> </system.web> </configuration> and this is the resulting response header for http://cdn.test.dev/css/global.css: Cache-Control: private,public, max-age=31536000 Content-Type: text/css Content-Encoding: gzip Expires: Wed, 01 Jan 2020 00:00:00 GMT Last-Modified: Mon, 06 Sep 2010 08:53:06 GMT Accept-Ranges: bytes Etag: "0454eca04dcb1:0" Vary: Accept-Encoding Server: Microsoft-IIS/7.5 X-Powered-By: ASP.NET Date: Mon, 06 Sep 2010 14:57:08 GMT Content-Length: 4495

    Read the article

  • OWA (Outlook Web Access) won't give user a link to attached document. Gives "Open as Web Page"

    - by Jerry Mayers
    I have been testing OWA and have found that when there is an attached file on a message it shows the "Attachments:" then the file name (but no link) and then [Open as Web Page] link. The user has to click on that which opens a new window and previews the document, then the user gets "You are currently viewing:" followed by the file name. This time the file name is a link that they can open/download. This happens for Office documents and PDF file at least. How do I make it so users can get a direct link to the file on the original message window?

    Read the article

  • Looking for Full Screen Web-Browser in Vizio or Samsung TV's

    - by ScottCate
    I see that Google TV, inside Sony, has a Chrome browser. Is the same thing possible inside Vizio, or Samsung, or any other TV? The 42" Vizio is $499 at Costco, and the Sony with Google TV is $999. We're looking at a buch of these TV's that have build in Wi-Fi, to place around the office as Big Dashboards, and I don't want to have a computer attached to get web content on the screen. It's my understanding that both Samsung Apps, and Vizio Apps, use Yahoo Widgets as their ap engine. Anyone know of a way to get a full screen browser going, without attaching a computer? We thought of attaching an Apple TV, which can be jailbroken, but that is just another piece of equipment to lose, break, etc. Thank you!

    Read the article

  • Dynamically loading chef recipies from URLs

    - by andy
    I'm deploying a web app on AWS. I intend to use chef to build AMIs which I'll then put into production. I want to have Chef monitor a URL stored in simpleDB. The URL would point to a tarball in S3. There would be different URLs, one for a config tarball, one for a code tarball. When I update the URL in simpleDB, I want chef to spot this and pull in and apply the configs/deploy the code. Is this possible? Has anything like this been done before or would I need to roll my own code? I think Chef can monitor URLs, but how would be the best way of getting it to load that URL from simpleDB?

    Read the article

  • Changing Administrator password on a Windows 2008 web server

    - by Nick
    I've just taken over the administration of a Windows 2008 web server from a previous employee on a temporary basis. I need to change the Admin password as soon as I can but I've noticed that quite a few of the services also run under this account. So: Is there a quick way to find out which services will be affected by me changing the password or is it a question of going down the list? It doesn't seem right to me that the Admin account is used in this manner; should I create a different account for these services, or is using the Admin a/c standard practice? I realize everyone's servers / networks are set up differently, but are there any other items I should be aware of when changing the Admin password? Thanks

    Read the article

  • Trouble migrating Joomla from Linux to Windows server

    - by Matt
    I'm having some trouble migrating a Joomla website from a Linux server to a Windows server. The database came over fine, Besides that, all I've done is download all the files from the current site, and change configuration.php so the log and tmp directories show "./tmp/" and "./logs/" I keep having an error in the PHP log stating PHP Fatal error: Class 'JTable' not found I've downloaded the site multiple times now, and I'm convinced this is a configuration problem, and not a missing file problem. I've even tried installing a mod for backup on the linux box to try and migrate the site, but sadly the mod had problems installing. The new server is running IIS 7.5 on Windows Web Server. PHP 5.2.14 and MySQL

    Read the article

  • Vmware Workstation Development Server on Laptop

    - by Koobz
    I'm running an Ubuntu guest OS on my Windows 7 laptop. Currently, I have it set up for bridged networking. The guest is os is configured for a static ip of 192.168.1.115, which depending on the network I'm connected to, may not be available. When I want to view my development work I hit that up in my web browser. I'm really looking for the following scenario: 1) My guest OS ip address stays constant. 2) I can access my guest os even if I don't have an internet connection, or a lan/router. 3) I can share files with my guest/host. How does one accomplish this using VMware Workstation?

    Read the article

< Previous Page | 399 400 401 402 403 404 405 406 407 408 409 410  | Next Page >