Search Results

Search found 32343 results on 1294 pages for 'good practice'.

Page 217/1294 | < Previous Page | 213 214 215 216 217 218 219 220 221 222 223 224  | Next Page >

  • User not able to change permissions on folder in network share for Windows Server 2003

    - by Ishmael
    I have a user that has rights to change permissions within a certain folder on a Windows Server 2003 network share. Although I grant this person the rights to change permission and ownership this person cannot change any permissions. The network share is also set to allow everyone full control. (I understand it is bad practice yet this is how it was first configured) Is there any reason why this domain user cannot modify permissions on the folder and files they create?

    Read the article

  • Is adding users to the group www-data safe on Debian?

    - by John
    Many PHP applications do self-configuration and self-updating. This requires apache to have write access to the PHP files. While chgrp'ing them all to www-data appears like a good practice to avoid making them world writable, I also wish to allow users to create new files and edit existing one. Is adding users to the group www-data safe on Debian? For example: 775 root www-data /var/www 644 john www-data /var/www/johns_php_application.php 660 john www-data /var/www/johns_php_applications_configuration_file

    Read the article

  • LDAP RBAC model

    - by typo
    Hi does anybody can tell me about best practice to model RBAC on LDAP ? I'm very confused, not sure if I should think about LDAP groups as role, or just user in some custom OU. Any real-life examples with tasks-operations\roles\user scheme (one user, multiple roles per user, multiple operations-tasks per role) ? BTW:Target systems are .net, java and iSeries

    Read the article

  • Why does squid reject this multipart-form-data POST from curl?

    - by keturn
    This fails: $ curl --trace multipart-fail.log -F "source={}" http://127.0.0.1:3003/jslint With a squid status 417 error, ERR_INVALID_REQ. trace of failing curl request trace of successful curl request that uses urlencoding (curl -d) instead of multipart (curl -F) formatted version of squid's error message I've never had this in practice through a web browser, so it's probably curl usage instead of squid, but if I tell curl not to use the squid proxy, the web application on the other end accepts it just fine. (If there's a more appropriate StackExchange site for this, please let me know.)

    Read the article

  • Emacs and windows manager keyboard shortcuts without Win key

    - by Little Bobby Tables
    I found a classic M-Series keyboard and I want to use it. However, it does not have the "Windows" key (a.k.a "Super"), only the Shift, Control and Alt modifiers. My keyboard shortcuts are cluttered as-is, since that I try to control both Emacs and the windows manager (Gnome) only from the keyboard. I rely on the "Super" key to identify the windows manager shortcuts. What it the best practice for keyboard-centric work without the "Super" key?

    Read the article

  • Is there a performance penalty using in-place models/families in a large Revit project

    - by Jaips
    (I'm quite new at Revit so apologies if my concepts are a bit inaccurate) I have heard using, in-place models in Revit projects is poor practice since it can slow down a large project. However I noticed Revit also organising inplace models lumping them with the rest of the families. So my question is: Is there really any performance penalty/benefit to be had by inserting families from an external file as opposed to creating inplace models in a Revit project?

    Read the article

  • Where to put my backup.sh?

    - by Temnovit
    I'm writing a shell script that will make backups of my system ( like this: https://help.ubuntu.com/8.04/serverguide/C/backup-shellscripts.html ). What is the best location in my system, to store this file? I know, I can put it anywhere, but what will be if it will be stored in a directory being backed up? What is the best practice here? I'm running an Ubuntu-server.

    Read the article

  • Cheap web hosting that supports php/mysql

    - by Cortopasta
    Does anyone know where I can find web hosting that supports php/mysql? I'm just using it to practice updating/changing live web sites, and to run code in an environment other than my XAMP stack on my thumb drive. It doesn't need to be fast or able to use a lot of bandwith, just somewhere I can play with code to sharpen my skill set

    Read the article

  • Is there a performance pentalty using in-place models/families in a large Revit project

    - by Jaips
    (I'm quite new at Revit so apologies if my concepts are a bit inaccurate) I have heard using, in-place models in Revit projects is poor practice since it can slow down a large project. However i noticed Revit also organising inplace models lumping them with the rest of the families. So my question is: Is there really any performance penalty/benefit to be had by inserting families from an external file as opposed to creating inplace models in a Revit project?

    Read the article

  • Port-forwarding HTTPS web server

    - by James Moore
    I have port forwarded our front-facing IP to an internal HTTPS server. The browser does not connect. A wget command determines that the certificate is self-signed for the internal IP. Hence why the browser is refusing to display the page properly. What is the best-practice scenario for this sort of stuff? Thanks

    Read the article

  • Any hardware/software routers that support Full Cone NAT?

    - by Ian Boyd
    i'm trying to get Teredo to function on my machine. Most routers, it seems, refuse to forward packets from any host other than the one i specifically connected to first. Teredo requires full Cone NAT in order to function. Does any router, hardware or software, allow full cone NAT? Is this an oversight by the designers of Teredo that nobody, in practice, can use it? i've tried m0n0wall pfsense D-Link Linksys SMC

    Read the article

  • Local or public NTP servers?

    - by BeeOnRope
    For a relatively large network (thousands of hosts) - what are the arguments for and against running a locally managed (pool of) NTP server(s) (perhaps periodically set via some public NTP server) and having all other hosts on the network use that (pool of) NTP server(s) versus having all hosts simply use public NTP servers directly, say via ntp.pool.org? Aside from the pros and cons, What is typical best practice today?

    Read the article

  • how to change zone of a extented webapplication-SharePoint?

    - by Ryan
    I have extended a webapplication and set its zone to Intranet but now we decided to make that site anonymous I made all the settings in authentication provider and site collection settings but its still showing its zone as Intranet . I read that its best practice to keep Internet as the zone for anonymous access.. How can I change it now and does it affect leaving it as Intranet?

    Read the article

  • Recommended Practices for Managing httpd.conf for different environments

    - by James Kingsbery
    We have several different environments (developer dekstop, integration, QA, prod) which should have slightly different variants of the same httpd.conf file. As an example, the httpd.conf file configures that httpd should act as a reverse proxy and proxy certain URLs to Jetty, but the hostname of the Jetty instance is different in each environment. Is there a recommended practice for mangaging these kinds of differences? I looked around the apache documentation for the httpd.conf file and I didn't see anything that does what I need.

    Read the article

  • SQL Connection String to access localhost\SQLEXPRESS

    - by user34683
    I've installed SQL Express on my PC hoping to do some practice creating tables and then modifying them. I coded a webpage in Visual Studio to, basically, SELECT * from a table in the SQLEXPRESS, but I can never get the connection string to work. Please help My connection string "Data Source=localhost\SQLEXPRESS;Initial Catalog=test;User Id=xaa9-PC\xaa9;Password=abcd;" Error Message: Query is select * from tblCustomers where username='johndoe' error is Login failed for user 'xaa9-PC\xaa9'.

    Read the article

  • How many megabytes per second may I expect from a gigabit USB 2.0 network card under linux?

    - by Nakedible
    I'm interested in the actual real-word throughput attainable with an external 1000BaseT USB 2.0 network card under Linux. I have been able to attain 90 megabytes per second on a PCI-E interface, but the USB 2.0 bus has a theoretical limit of 480Mbit/s, and in practice less than 40 megabytes per second. Is the actual throughput attainable with such a card under linux 40, 30, 20, or even as low as 10 megabytes per second, eg. no better than a normal 100BaseT network card?

    Read the article

  • Oracle Virtualbox on statically compiled kernel

    - by aking1012
    I can't seem to find any documentation on the subject. I'm working on putting together a linux install for a fairly "dirty" environment. Best practice there would be a statically compiled kernel with no module support. I can already do the customizations to strip out unnecessary drivers/etc to get the performance and disable module support. Does anyone have a link or any ideas on how to get the Oracle Virtualbox module (not the OSE one, I need USB passthrough) compiled in?

    Read the article

  • Avoiding QoS degradation for video streaming clients

    - by aarege31
    Suppose I have two routers connected via a 1Gbit connection. A client behind router 1 streams to a client behind router 2 while other clients behind router 1 transmit data to other clients behind router 2. Are there any best practice policing, scheduling or queue management algorithms available that help a beginner understand what is necessary to prevent QoS degration in simple cases as above as well as in real world environments?

    Read the article

  • Resource Pools and VMTools Best Practices

    - by LucD
    Are there any published (or non-published) best practices with using Resource Pools in a View environment? Using Shares/Limits etc...Also, when installing the VMTools on a desktop, are there any best practice configurations with running the tools within a desktop?Resource Pools and VMTools Best Practices

    Read the article

  • 20GB+ worth of emails in my /home what is a better solution for that?

    - by Skinkie
    My email storage requirements are outgrowing anything reasonable with respect to local mail storage. As we speak 99% of my home partition is filled with personal mail in Thunderbirds mail dirs. Needless to say, this is just painful, badly searchable and as history has proven me that backups work, but Thunderbird is capable of loosing a lot of mail very easily. Currently I have an remote IMAPS server (Dovecot) running for my daily mail, accessible from anywhere, which from my own practice works efficiently up to about 1000 emails. Then some archive directories should be used to move mail around. I have been looking into DBMail, but I wonder if I make my case worse or better which such solution. None of the supported database employ string deduplication or string compression out of the box, so is this going to help me with 20GB+ mail? What about falling back to a plain old IMAP server? A filesystem like ZFS would support stuff like GZIP transparently, which could help. Could someone share their thoughts? The 20GB mostly consists of mailinglists, and normal mail. Not things like attachments. To add some clarifications; As we speak, my mail is not server side indexed at all - only my new mail arrives at a remote IMAP server. It is all local storage from former POP3 accounts, local mirrored Gmail and IMAP accounts. In my perspective it is not Thunderbird that sucks, its fileformat that sucks. Regarding the 1000 mails. On the road I am using Alpine and MobileMail, quite happy with both of them, but some management is required to actually manage the mail. Sieve helps a lot with that, but browing through 10.000 e-mails is not fun, especially not on a mobile client. I am quite happy with Dovecot, never had any issues with it. I just wonder if this is the way to go. Or if there are any other better solutions. What my question is: what is the best practice solution that allows 20GB+ mails and is -on demand remotely accessible, easy to backup and archive worthy. It doesn't need to be available 24x7. The final approach I took was installing a local IMAP server (Dovecot), configured it for being my archive, using the following guide: http://en.gentoo-wiki.com/wiki/Dovecot/InstallThunderbird

    Read the article

  • How to use Git over multiple similar systems

    - by Spidfire
    I have a system I need to duplicate over several systems and make minor changes like change less/css variables and configuration files. Is there a best practice for these kind of problems? I currently do: git clone repo cp ../default/config.js config.js ... for several files or should I create different branches of the same repo or should I create an repo for the changes? It is currently doable but it will get annoying if I get more than 5 similar systems.

    Read the article

  • what permissions are granted to a sql server database owner?

    - by Charles Hepner
    I have been trying to find out what permissions are granted to the owner of a database in SQL Server 2005 or higher. I have seen best practices questions like this one: What is the best practice for the database owner in SQL Server 2005? but I haven't been able to find anything specifically addressing what the purpose of having a database owner in SQL Server is and what permissions are granted as a result of making a given login a database owner. If the owner of the database is disabled, what would stop working?

    Read the article

< Previous Page | 213 214 215 216 217 218 219 220 221 222 223 224  | Next Page >