Search Results

Search found 21283 results on 852 pages for 'control flow'.

Page 585/852 | < Previous Page | 581 582 583 584 585 586 587 588 589 590 591 592  | Next Page >

  • How do I create email addresses for all my users using my domain ([email protected])?

    - by user33012
    My client wants to create email addresses for all their dotnetnuke users using their domain. The point is to keep the user's email addresses 'private' while still allowing communication through a public email address that they can control. It's not necessary to have a full webmail interface (although that would be nice). I'm thinking it would be enough just to forward any mail on and just act as a gateway. So if an email was sent to [email protected], it would be forwarded on to the email address associated with the dotnetnuke account with username 'rwain'. Is this possible to do in a shared hosting environment? Or do I need to create some custom mail server that does a conversion of the email address and forwards it?

    Read the article

  • selecting files in windows-7 explorer just by hovering

    - by bortao
    One of the biggest annoyances for me of Windows 7 vs Windows XP is that you cant select files properly by hovering over them on Windows Explorer (or multi select by holding ctrl). When i hover an item it gets light blue, not selected. Only after a time or after more hovering, it gets selected. I set the delay (MouseHoverTime) to 1. Windows seems to avoid wrong selection by requiring an additional time or space of mouse hovering before actually selecting the item. I tried playing with the values MouseHoverHeight, MouseHoverWidth, MouseSensitivity, MouseThreshold1 and MouseThreshold2 in HKEY_CURRENT_USER\Control Panel\Mouse, but they seem to have no effect over this. (they are currently on (20, 20, 16, 6, 10, respectively). What i want: to immediately select an item whenever the mouse gets over its bounds.

    Read the article

  • Windows 7 doesn't connect to mapped drives at start up.....

    - by danbo
    We are testing Windows 7 (32bit version) and logon to a windows 2003 domain that runs a logon script whichs maps our drive letters. We have no control over the domain. Of our 7 test users, 5 continually recieve red X's on their network drives after they logon. Double clicking on them connects the drives, however, any aplpication that requires files on the fileserver will fail since it thinks it has no connection to the drive. We have tried several reg edits (Enable Linked connections, KeepConn), we have tried to find information in the event viewer to no avail. We have also looked at any differences in NIC driver versions (none). The other 2 computers that can connect without problem are local admins, but, if we logon to the 5 that have the problem with the local admin credentials we get the red x's as well??? This one is a real head scratcher......

    Read the article

  • Windows 7 Boot Loader - Remove 30 second waiting time

    - by derekhh
    I've installed Fedora 17 and Windows 7 on two different hard disks as a dual-boot system. The default boot loader is GRUB 2 maintained and configured in Fedora 17. When I startup and choose "Windows 7 bootloader (on /dev/sda1)" in GRUB 2, Windows 7 boot loader will appear, with only Windows 7 as the only choice and also the default operating system, with a 30 second waiting time if no input is detected. I'm trying to see if it is possible to remove this 30-second waiting time. I've tried to follow the instructions on the Web by configuring default operating system in the control panel but seems there is something wrong with it. I've also tried to use "bcdedit enum /all" but still receives error prompts. Update: Added my boot tab screen in msconfig.

    Read the article

  • How to configure and access IIS in windows 7

    - by user28233
    I've installed IIS on Windows 7(Ultimate) by going to control paneladd/remove features/ then checking the IIS. How do I access something like a preference editor like in wamp server. In wamp server I go to http://localhost/phpmyadmin. And accessing the pages I made in http://localhost/pages/ What about in IIS? How do I access something like phpmyadmin. And how do I access the webpages that I made. This is only for educational purposes. And I do not need a site to be accessible on the net. And I already have something like this: What do I do with it. Please help,thanks

    Read the article

  • what to use instead of laptop-mode?

    - by playcat
    hello, i have ubuntu 10.10 64bit on hp 6735s (turion processor). it overheats, and i'm forced to use turion power control in order to keep core temperature to a reasonable level. one more measure that i use is putting my processors to conservative mode. that way, i'm perfectly happy with its performance, and heat is where it should be. however, after my latest upgrade, something happened - cores are back to ondemand by default, and i'm not sure if turionpowercontrol is working any more (ps axu | grep urion shows no process). in addition, i read somewhere that laptop-mode uses hdd spindown for preserving data/energy, and that hdds have only a limited amount of those spindowns, so laptop-mode usage can actually shorten the life of my hdd. i'm wondering if there is a good way to set my cores to automatically go to conservative mode? also, what's the good way to see what is the voltage my cores use? on windows i use cpuz tools. thx and sorry for the long explanation.

    Read the article

  • Can't delete folder and I am admin. "You need permission to perform this action. You require permission from..."

    - by Tony_Henrich
    I can't delete a certain folder in windows 7. It's not a system folder. It was created by an app. I get an error message "You need permission to perform this action. You require permission from Tony....". I am an admin and logged in as myself Tony. I have full ownership of the folder with full control permission. I rebooted the machine. I killed the app which created the folder. I am out of ideas. Why is Windows 7 not letting me delete the folder?

    Read the article

  • Server goes offline. What to look for?

    - by Jonathan Sampson
    I'm using a new virtual server through GoDaddy, and this morning I received a call from the powers that be informing me our website was offline. After confirming this, I requested a power cycle through our GoDaddy control panel, and within a minute or two the server was back online. I made the call, and reported the news that we're back up. Of course, a couple minutes later we're down again. I tried connecting through PuTTy, and it takes forever to prompt me for a username, and each successive prompt takes a long time to come up. I'm using CentOS. So my questions are: How can I determine the cause? What types of things can I do to prevent this in the future? One interesting, and perhaps relevant, observation is that yesterday our bandwidth consumption was about 20% greater than our top figures from the past month.

    Read the article

  • Running OpenVZ virtual servers within a Xen XCP vritual server? Bad practice?

    - by Damainman
    I have a 1 server with 8GB RAM and 2xQuadcore Processors. It currently has the Xen XCP installed on it, and centos6.2 x64 running on a virtual machine. I have a server control panel software that I want to use and it allows the administration via a web interface for Openvz machines. My questions are: Would this be considered bad practice? Would there be a big performance hit? Should I avoid this all together or am I going about it all wrong? Thank you in advance.

    Read the article

  • Missing 16:10 resolutions with Nvidia drivers (Can't add resolutions)

    - by Wuinny
    I have a laptop with a Nvidia 9650M GT and used the drivers that Seven brought me. It works fine but Metro 2033 tells me that I have to upgrade my drivers to play the game. So I did it. But since I did a clean install of the new Nvidia drivers, I just have 1440*900 or 4:3 resolutions. I usually played with 1280*800 or 1184*740 (for performance issue) With the "old" drivers I was able to create custom resolution (1184*740) in Nvidia control panel but now when I try it tells me that "my monitor cannot support this resolution". When I insist, it works but soon as I shut down my computer I have to recreate it.. Do anyone have a fix?

    Read the article

  • Cannot usermod -L in LightDM scripts

    - by user95723
    I'm running Xubuntu 12.04 and use the LightDM. I want to restrict access to the machine as a kind of parental control. This is how it should work I hook in a script that executes just before the greeter comes up. Within that script some awk processing will read an entry in a config file and will trigger a usermod -L or usermod -U depending on whether the user is allowed to login. While user is logged, a cron job will count down the entry in the config and forces a xfce4-session-logout if time is up. A cron job running on a server will upload the "credits" on a daily base. How is this idea? That's theory, now for the problems It appears for some unknown reason, the usermod command is not executed, neither as part of a display-setup-script nor within the greeter-setup-script. I wrote a small sandbox script usermod -L johndoe 2error.txt touch /etc/blabla 2error.txt The script is executing, cause the blabla file is existing. That means that the script must have been executed with root privileges. error.txt is empty but the usermod command has just no effect. Is this a bug or a feature. What's wrong? Best regards and thank you Oli

    Read the article

  • How to Increase Memory Allocated to IIS .NET Application?

    - by Mark Hansen
    We are using Windows 2008 R2 and IIS 7 running on Amazon EC2. IIS is running a single .NET application written in C#. We are having performance issues and I want to give the application more memory, but I cannot figure out how to do it. How do I control the amount of memory that the CLR gets? I'm a total newbie with IIS, .NET and the CLR. If I were working with Java, I would just use the -Xmx flag to increase the memory available to the JVM (e.g., -Xmx3000m for 3GB). But, I cannot seem to figure out how to do this in the Windows world.

    Read the article

  • Why is text on my TV monitor all grainy?

    - by searchfgold6789
    I have successfully installed the Catalyst drivers and hooked up my 1080p TV to the Radeon HD 5750's HDMI port and now it is working. There only seems to be one issue - most images appear OK, at least to my eyes, but text is grainy and difficult to read (not sure how well it can be seen in the picture): It works fine with a smaller screen, and hopefully you can see it with the new pictures which are only slightly better, but images are rendered a little funny too. Changing antialiasing does nothing and changing the refresh rate only makes a slight change if any. The small monitor I have works fine. Smaller resolutions such as 1024x768 look OK but are the wrong aspect ratio. I have tried renaming the input mode on the TV to "PC", and that helped... a little bit. Also, I tried a different PC with the same result. Messing around extensively with the ATI Catalyst Control Center also did not help. I am wondering if there is anything else I can try to increase the quality so it is more usable.

    Read the article

  • Do you have to recreate workspaces after upgrading a TFS 2008 server to TFS 2010?

    - by Clara Oscura
    I am just reposting this thread from a MSDN forum since it seems to be unavailable. It was very useful when I was having trouble with my folder mappings after migrating to TFS 2010. Question: I opened VS2008 and connected it to the upgraded 2010 TFS server.  Upon clicking any of our Team Projects in source control explorer I get "Team Foundation Error - The workspace MYWORKSPACE;DOMAIN\MYUsername already exists on computer MYPCNAME." Answer: The same local paths on your machine are mapped to 2 different workspaces, one on the preupgrade server and one on the postupgrade server.  It's not safe to have multiple workspaces on different servers mapped to the same local paths b/c you could pend some changes while connected to one server, and the other server would have no idea what you did.  You should either delete your conflicting workspaces from one of the servers (if you don't need them on both), or test the new TFS instance from a new workspace (on different machine). If you want to test an existing production workspace on both servers, then yes, you will have to mess around with the workspace cache. You don’t have to delete the entire cache, you just need to run "tf workspaces /remove:* /server:<serverurl>" to clear the cached workspaces from a server (the command won't delete the workspaces), and possibly "tf workspaces /server:<server>" to refresh the workspace cache for a given server.  You will also have to do back up and restore the workspace before switching servers or your local files could be inconsistent. From the “Microsoft Visual Studio Team Foundation Server 2010 Beta 1” forum (not available anymore?) Technorati Tags: TFS 2010,TFS Workspaces,Team System,Team Foundation Server 2010

    Read the article

  • How to move from Programmer to Project Lead

    - by DoctaStooge
    At my job, I'm currently a programmer, but in the next few weeks I'll be taking control my own project. I was wondering if anyone else here has been in the same situation, and if so, what advice you can offer to help me be able to better run my project. Experience in dealing with contractors would be greatly appreciated. A little more info: Project will have 3 people including myself, with extra people coming in when needing testing. The project has been programmed mainly by 2 people I would like to contribute to the programming as I like doing it and think I can add to the program, but am afraid of how the contractors will react. I don't want to create bad feelings which may harm the project. EDIT: Forgot to mention that I'll have to be picking up communications with customers to make sure their needs are met. Any advice on talking to customers cold would be greatly appreciated. EDIT 2: This is not a new project, I'm picking it up around version 6. Sorry that I didn't make it clear before.

    Read the article

  • Unauthorized access error to html pages in IIS 7.0

    - by George2
    I am using VSTS 2008 + C# + .Net 3.5 + IIS 7.0. I have created a new web site and put an html file into the directory. And when I use browse function in IIS manager to browse the html file, I met with the following error, any ideas what is wrong? BTW: I am very confused about unauthorized error since I run the worker process under administrator account. From the error message, I am confused why the logon method is anonymous and not using administrator account? HTTP Error 401.3 - Unauthorized You do not have permission to view this directory or page because of the access control list (ACL) configuration or encryption settings for this resource on the Web server. Module IIS Web Core Notification AuthenticateRequest Handler StaticFile Error Code 0x80070005 Requested URL http://localhost:80/a.html Physical Path C:\test\simplehosttest\a.html Logon Method Anonymous Logon User Anonymous thanks in advance, George

    Read the article

  • How can I delete Time Machine files using the commandline

    - by Tim
    I want to delete some files/directories from my Time Machine Partition using rm, but am unable to do so. I'm pretty sure the problem is related to some sort of access control extended attributes on files in the backup, but do not know how to override/disable them in order to get rm to work. An example of the error I'm getting is: % sudo rm -rf Backups.backupdb/MacBook/Latest/MacBook/somedir rm: Backups.backupdb/MacBook/Latest/MacBook/somedir: Directory not empty rm: Backups.backupdb/MacBook/Latest/MacBook/somedir/somefile: Operation not permitted There are a number of reasons I do not want to use either the Time Machine GUI or Finder for this. If possible, I'd like to be able to maintain the extended protection for all other files (I'd like not to disable them globally, unless I can re-enable once I've done my work).

    Read the article

  • mass deploy Oracle patches with OEM with different OS's

    - by bobsmith12
    I have not been able to get this to work. We are running our OEM grid control database/oms on Red Hat 5(32 bit), but our databases are on Solaris x86-64. I could not mass deploy agents since the Operating Systems were different. When I download patches it is by OS. Is there a way to mass deploy to multiple operating systems? I have alot of databases. I was given the redhat server for OEM because it was available. We are have 10.1,10.2, and 11.1 databases. OEM DB is 10.2.0.5

    Read the article

  • "Cannot allocate memory " error whle copying data from window to ubuntu

    - by John
    I have Ubuntu 9.10 installed inside VM of server 2008. WHen i try to copy the data from the network and paste insid ethe Ubuntu it says error called "Cannot allocate memory " I have 3GB RAM attached to the Ubuntu I tried above suggestion but still im unbale to copy file from my host machine i.e. Windows XP to my Ubuntu machine ( which is at Virtual Machine) Im trying to copy jdk-1_5_0_22-linux-i586.bin file whose size is 47.4 MB Is there any other work around for this problem???? I tried Set the following registry key to ’1': HKLM\SYSTEM\CurrentControlSet\Control\Session Manager\Memory Management\LargeSystemCache and set the following registry key to ’3': HKLM\SYSTEM\CurrentControlSet\Services\LanmanServer\Parameters\Size but still im unbale to copy file from my host machine i.e. Windows XP to my Ubuntu machine ( which is at Virtual Machine) Im trying to copy jdk-1_5_0_22-linux-i586.bin file whose size is 47.4 MB Is there any other work around for this problem????

    Read the article

  • How do web servers enforce the same-origin policy?

    - by BBnyc
    I'm diving deeper into developing RESTful APIs and have so far worked with a few different frameworks to achieve this. Of course I've run into the same-origin policy, and now I'm wondering how web servers (rather than web browsers) enforce it. From what I understand, some enforcing seems to happen on the browser's end (e.g., honoring a Access-Control-Allow-Origin header received from a server). But what about the server? For example, let's say a web server is hosting a Javascript web app that accesses an API, also hosted on that server. I assume that server would enforce the same-origin policy --- so that only the javascript that is hosted on that server would be allowed to access the API. This would prevent someone else from writing a javascript client for that API and hosting it on another site, right? So how would a web server be able to stop a malicious client that would try to make AJAX requests to its api endpoints while claiming to be running javascript that originated from that same web server? What's the way most popular servers (Apache, nginx) protect against this kind of attack? Or is my understanding of this somehow off the mark? Or is the cross-origin policy only enforced on the client end?

    Read the article

  • Moving to an outside Exchange provider - How to create new Outlook profile on multiple PCs?

    - by imagodei
    Following an acquisition we are switching from our local Exchange to an outside provider. I need to change account settings / create new profile so that users are able to access their new mailbox, and I hate the idea of doing it manually on every single client. I have tried quick and dirty solution with copying profile information from Registry on one computer to the other, hoping that I could fix user info later. Quite obviously, this didn't work: Outlook recognized the profile, I could select it at the application startup, but there the fun ends - the profile was corrupt. I couldn't change settings in Mail settings on Control panel, either. Didn't find any Group policy solution. Is there any application, script or other solution, that would help me save time commuting from one PC to the other and typing the same info over and over again?

    Read the article

  • $DISPLAY dependent gtk themes

    - by Vlad Seghete
    I have a computer at home that I log into remotely. The "monitor" for it is a TV, so I want gtk applications to use a large font and icon theme, which I managed to do by editing the ~/.gtkrc-2.0 file and some other similar stuff. What I want to be able to do is have a separate theme for when I'm logging in remotely. The best way to explain is that I would like my gtk theme choice to be dependent on the X display that the application is started on. For example, if I start something on :0.0 then that is the TV and I want large fonts, but if I start it on localhost:10.0 I want to use a regular size font, because it will get rendered on my laptop screen. The elegant solution would be to have some sort of IF statement in the .gtkrc-2.0 file that checks the $DISPLAY variable and behaves accordingly. The problem is I can't find any documentation on control structures in .gktrc files, or if it's even possible to do that.

    Read the article

  • Programmatically change height of the window more than the height of the screen in Mac OS X

    - by nezlooy
    Respected low-level users of Mac OS, please, help. I'm trying programmatically to change height of window of safari (or other window). I'm was trying used AppleScript and AXUIElementSetAttributeValue of Carbon, but none of these methods can't increase the window bigger than the height of the screen. But, the width changes without any problems. I guess I'd be starting with something like SetWindowPos with SWP_NOSENDCHANGING flag under Win32. Maybe, you can disable this functionality, which monitors the size of the window or completely shutdown the one who is responsible for it? Note that I want to do this programmatically from an external process - I'm not asking how to control just my own app's window size and position. Thanks.

    Read the article

  • Users can benefit from Session Tracking

    I use to work for a large Dental Plan marketing website a few years ago and they had a large customer-driven website that sold Dental Plans to consumers. Their website started tracking users as soon as they hit their web servers, and then they logged everything they could about the user. There are a lot of benefits for using session tracking for both the user and the website. Users can benefit from session tracking due to the fact that a website can retain pertaining information for the user so that they do not have to re-enter the same information repeatedly. In addition, websites can hold specific items in a cart for each user so that they can pay for all of their  items at once when they are ready to complete their purchases. Websites can also benefit from session tracking because they can determine where a specific user came from and which advertising partner gave them a sale. This information is very useful when deciding on where to spend an advertising budget. There is only one real disadvantage when it comes to session tracking, Users can not really control what is actually tracked by a website. Yes, they can disable cookies and this will help, but that means that no tracking can be done at all. Most sites require users to have cookies enabled in order for users to make purchases or login to their accounts.

    Read the article

  • How to emulate a domain name - Webmin Setup

    - by theonlylos
    I am currently working on a client project where they are using a custom CMS which relies on having the specific domains configured for it to work properly. So in English, that means that when I try running the site on my test environment, the entire website fails because it isn't located on the primary domain (and I'm pretty sure the domain is hard coded since there's no control panel to adjust the file locations). Anyway what I wanted to ask is whether it is possible to use my test environment URL but have Apache and the DNS emulate my clients website URL locally, rather than calling the actual name servers. Right now I have a virtual host setup in Apache but I am not sure where to go from there. Any assistance is greatly appreciated.

    Read the article

< Previous Page | 581 582 583 584 585 586 587 588 589 590 591 592  | Next Page >