Search Results

Search found 11568 results on 463 pages for 'config spec'.

Page 209/463 | < Previous Page | 205 206 207 208 209 210 211 212 213 214 215 216  | Next Page >

  • Wordpress plugin installation error.

    - by Steve
    I'm trying to upload secure-wordpress.1.0.6, and I receive the following error: Warning: touch() [function.touch]: open_basedir restriction in effect. File(/abs_path/wordpress/tmp/secure-wordpress.tmp) is not within the allowed path(s): (/abs_path/:/abs_path/:/usr/local/lib/php:/tmp/php_upload) in /abs_path/public/www/wordpress/wp-admin/includes/file.php on line 199 Download failed. Could not create Temporary file. The /wp-content folder and all it's subfolders have 777 permission. I've added the following two lines to wp-config.php: putenv('TMPDIR='.ini_get('upload_tmp_dir') ); define('WP_TEMP_DIR', ABSPATH .'wp-content/uploads/'); What else should I try? I am using Wordpress 3.04 in a PHP 4.49 environment.

    Read the article

  • Software Tuned to Humanity

    - by Phil Factor
    I learned a great deal from a cynical old programmer who once told me that the ideal length of time for a compiler to do its work was the same time it took to roll a cigarette. For development work, this is oh so true. After intently looking at the editing window for an hour or so, it was a relief to look up, stretch, focus the eyes on something else, and roll the possibly-metaphorical cigarette. This was software tuned to humanity. Likewise, a user’s perception of the “ideal” time that an application will take to move from frame to frame, to retrieve information, or to process their input has remained remarkably static for about thirty years, at around 200 ms. Anything else appears, and always has, to be either fast or slow. This could explain why commercial applications, unlike games, simulations and communications, aren’t noticeably faster now than they were when I started programming in the Seventies. Sure, they do a great deal more, but the SLAs that I negotiated in the 1980s for application performance are very similar to what they are nowadays. To prove to myself that this wasn’t just some rose-tinted misperception on my part, I cranked up a Z80-based Jonos CP/M machine (1985) in the roof-space. Within 20 seconds from cold, it had loaded Wordstar and I was ready to write. OK, I got it wrong: some things were faster 30 years ago. Sure, I’d now have had all sorts of animations, wizzy graphics, and other comforting features, but it seems a pity that we have used all that extra CPU and memory to increase the scope of what we develop, and the graphical prettiness, but not to speed the processes needed to complete a business procedure. Never mind the weight, the response time’s great! To achieve 200 ms response times on a Z80, or similar, performance considerations influenced everything one did as a developer. If it meant writing an entire application in assembly code, applying every smart algorithm, and shortcut imaginable to get the application to perform to spec, then so be it. As a result, I’m a dyed-in-the-wool performance freak and find it difficult to change my habits. Conversely, many developers now seem to feel quite differently. While all will acknowledge that performance is important, it’s no longer the virtue is once was, and other factors such as user-experience now take precedence. Am I wrong? If not, then perhaps we need a new school of development technique to rival Agile, dedicated once again to producing applications that smoke the rear wheels rather than pootle elegantly to the shops; that forgo skeuomorphism, cute animation, or architectural elegance in favor of the smell of hot rubber. I struggle to name an application I use that is truly notable for its blistering performance, and would dearly love one to do my everyday work – just as long as it doesn’t go faster than my brain.

    Read the article

  • System that splits passwords across two servers

    - by Burning the Codeigniter
    I stumbled upon this news article on BBC, RSA splits passwords in two to foil hackers' attacks tl;dr - a (randomized) password is split in half and is stored across two separate servers, to foil hackers that gained access to either server upon a security breach. Now the main question is, how would this kind of system would be made... codespeaking, for PHP which I commonly develop on my web applications, the database password is normally stored in a configuration file, i.e. config.php with the username and password, in that case it is understandable that the passwords can be stolen if the security was compromised. However when splitting and sending the other half to the other server, how would this go on when making a communication to the other server (keeping in mind with PHP) since the other server password would be stored in a configuration file, wouldn't it? In terms of security is to keep the other server password away from the main one, just exactly how would the main server communicate, without exposing any other password, apart from the first server. This certainly makes me think...

    Read the article

  • Super Key doesn't work on Kubuntu 12.04

    - by zigma80
    I have Kubuntu 12.04 installed on HP 430 Notebook PC. My problem is that when I press Super Key, nothing happened. Some of my Fn keys did not work either. I have changed the Keyboard Model in System SettingsHardwareInput DevicesKeyboard to Hewlett Packard Internet keyboard but doesn't seems to help. I have also Compiz-Config-Setting-Manager installed but I don't know how to make the right keyboard setting. I would like to have the Super Key launch Start Menu Launcher as in MS-Windows, is it possible ... Thanks for the help

    Read the article

  • How to install vmware tools?

    - by Tom
    I installed my Ubuntu in vmware, no I need install vmware tools, I got error: Searching for a valid kernel header path... The path "" is not valid. Would you like to change it?[yes] In CentOS, I run the following command to resolve this issue: yum install gcc-c++ yum install kernel-devel yum install kernel-headers yum -y update kernel But I don't know how to do in Ubuntu. Please help. Update I have tried the following command but nothing changed,still got error: Searching for a valid kernel header path... The path "" is not valid. Would you like to change it?[yes] sudo apt-get update sudo-get install build-essential linux-header-$(uname -r) sudo ./vmware-uninstall-tools.pl sudo ./vmware-config-tools.pl sudo ./vmware-install.pl Issue Changed: Run sudo ./vmware-uninstall-tools.pl, and delete the folder of /etc/vmware-tools then, run sudo ./vmware-install.pl Now I can successfully install vmware-tool.After restart, I can see folder of /mnt/hgfs, but can't see my shared folder.

    Read the article

  • SSIS 2008 Configuration Settings Handling Logic for Variables Visualized

    - by Compudicted
    There are many articles discussing the specifics of how the configuration settings are applied including the differences between SSIS 2005 and 2008 version implementations, however this topic keeps resurfacing on MSDN’s SSIS Forum. I thought it could be useful to cover the logic aspect visually. Below is a diagram explaining the basic flow of a variable setting for a case when no parent package is involved.   As you can see the run time stage ignores any command line flags for variables already set in the config file, I realize this is not stressed enough in many publications. Besides, another interesting fact is that the command line dtexec tool is case sensitive for the portion following the package keyword, I mean if you specify your flag to set a new value for a variable like dtexec /f Package.dtsx -set \package.variables[varPkgMyDate].value;02/01/2011 (notice the lover case v in .value) You will get errors. By capitalizing the keyword the package runs successfully.

    Read the article

  • Ubuntu Live USB: best practices for secure net traffic

    - by Och
    I want to to set up a live USB with Ubuntu, in the most secure way. So I want to have the persistent data on a second USB, something that its not that much problematic. How to configure a very safe Internet surfing (throughout a VPN?) Which are the best practices that could be implemented to have the Ubuntu live in a USB, the persistent data in other, and with the Internet access to a VPN (the Ubuntu privacy remix gives most of this, except the VPN config), Any ideas of how to combine the best of Ubuntu privacy remix, and Internet access to a VPN?

    Read the article

  • When Your ASP.NET MVC Custom Error Doesn't Render...

    You've put [HandleError] on your controller and you've set <customErrors mode="On"> in the web.config file. So why do you still see the yellow screen of death? I've heard various wrong explanations for this phenomena, including "it only works under IIS" and "it only works in release mode". But the custom error view does work in debug mode and it does work with the Visual Studio WebDev web server. I think the most common reason for the error view not to display is because the error view throws...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Allow any arguments for a given command with sudo

    - by Mark L
    I have the following sudo config entry which I added via sudo visudo: mark ALL = NOPASSWD: /usr/bin/lxc-ls* I can run lxc-ls with my user fine but I can't append any parameters without it demanding I prefix the command with sudo. $ whoami mark $ lxc-ls test-container $ lxc-ls --fancy lxc-ls: error: You must be root to access advanced container properties. Try running: sudo /usr/bin/lxc-ls Any idea how I can edit via sudo visudo to allow for any argument after the command? I don't want to prefix the command with sudo as I'm using a python library to execute the command and it's being funny about sudo prefixes.

    Read the article

  • When Your ASP.NET MVC Custom Error Doesn't Render...

    You've put [HandleError] on your controller and you've set <customErrors mode="On"> in the web.config file. So why do you still see the yellow screen of death? I've heard various wrong explanations for this phenomena, including "it only works under IIS" and "it only works in release mode". But the custom error view does work in debug mode and it does work with the Visual Studio WebDev web server. I think the most common reason for the error view not to display is because the error view throws...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • getting the user back where they came from with mod_form_auth

    - by bmargulies
    Using the mod_form_auth module in Apache HTTPD 2.4.3, I am looking for a way to have the user redirected to their original desired target after completing a login. That is, if I have a <Location /protected> ... form auth config here </Location> the user might browse to /protected/a, or to protected/b. In either case, they will be presented with the login form. However, as far as I can see, I must specific a single 'success' URL. I'm wondering if I'm missing some Apache feature that would allow me to, for example, cause the redirect to the login form go to something like: https://login.html?origTarget=/protected/a via some syntax on the AuthForLoginRequiredLocation statement?

    Read the article

  • Programmaticaly change lightdm autologin session type

    - by marcinpz
    I have problem with lightdm setting autologin session type (using my config tool). Lightdm logs into previously selected session. I see that autologin-session is UNIMPLEMENTED in version 1.2.1. Maybe I can somehow cheat lightdm? but it seems program writes actual state when stopping and overwrites my changes to /home/user/.dmrc and /var/cache/lightdm/user.dmrc. At the current I have two sessions: default and kiosk. in /etc/lightdm/lightdm.conf I have line: user-session=desktop but after I logged into kiosk session it starts all the time. Changing .dmrc and /var/cache/lightdm/dmrc/user.dmrc doesn't work.

    Read the article

  • Chrome dispay Glich when GPU acc. on

    - by user289172
    Hy Everyone! I have some graphics glich when enable gpu acc. in chrome. If i disable acceleraion it fix the glich but, in full HD most of flash videos frame rate drop above frogs a**, so very slow... :) Any idea, what can i do? I Upload 2 pics from glich. My hardware: CM Hyper TX 3 on AMD A5300 with HD 7480D 6 GB RAM I use official AMD driver from amd.com, Ubuntu 14.04(I upgraded from 12.04, in Ubuntu 12, same glitch with old drivers.) I reinstalld x.org completly and purge all config files, but this didnt resolve the problem. If i use other drivers, i have some other problem ie.:nor automatic resolution chnge when connect new monitor, false refreshrate for displays and etc. Thanks for the help! http://i.stack.imgur.com/9PkrK.jpg http://i.stack.imgur.com/AMS22.jpg

    Read the article

  • Quickly Application making it run on startup

    - by unknownone
    I have a PyGtk application that I made using Quickly, and I would like to have it run on startup when installed. How would I go about doing this? I'm not sure if sticking the .desktop file in ~/.config/autostart/ would make it work or not. If that will fix it, I don't know how to add it to that folder since Quickly packages the project for you and has it's own installation script. Is it possible to modify what it does on installation? If possible, I would also like to add the program in the System Settings Personal tab, but I do not know how to do that. Thanks

    Read the article

  • Dropbox icon dissapears right after login

    - by garvamel
    Even though Dropbox's app indicator dissapearing from the tray area seems like a recurrent enough problem, my issue is a litte different. When I login, I can see the app panel populating, and the dropbox icon does indeed appear (config'd as startup application), but after some other icons show up (bluetooth, battery, etc.) it's gone. It's still running though. I'm guessing it's having issues with staying pinned, and I don't know how to start addressing this problem. I have tried many if not all solutions provided here in the forums for the "icon missing" questions. So far: I've whitelisted everything regarding panel Uninstalled-reinstalled (with and without rebooting in between) Overwritten current installation Purged installation from terminal Installed from Software Center and from .deb file batch deleted every "dropbox" ocurrences from terminal (files and folders) and reinstalled Ran sudo apt-get install libappindicator1, it installed, but didn't solve anything I'm on Ubuntu 12.04 LTS - 64 bits. Any insight would me much appreciated!

    Read the article

  • Multiple Java EE Agents on Single Managed Server

    - by tina.wang
    A default JEE agent is created when you create domain, which is named as OracleDIAgent. 1. In Studio, duplicate the agent, change its name to genAgent, change the web application context to genagent. 2: Go to datasource of genAgent, drop all datasources.3: Generate server template. put the jar file under odi\common\templates\wls 4: Deploy this template by update the existing domain. Bring up the config.cmd, choose update existing domain. 5: Update the domain using the template that just generated. Go through the Configuration wizard. (I did not modify anything or configure anything here). 6: The wizard will give information says the deployment was successful. 7: Bring up the admin server and ODI_server1. 

    Read the article

  • Numlock doesn't work after logging in

    - by Tobi
    I've a strange problem after upgrading to Ubuntu 11.10 The Numlock of my Apple Keyboard (white, wired) doesn't work anymore. Strange thing because it works with the guest account but not with my user account. I did change the lightdm.conf to get the numlock working for the login and it works, but nor in my user account. Even stranger, pushing zero effects a right click? I also tried different keyboard layouts, but it shows Numlock on (stuck on) but doesn't work. Any idea which config file stores information about this?

    Read the article

  • 301 re-direct all external links to new domain

    - by Dean Legg
    I have changed the main domain to a sub-domain & would like to re-direct all external links to the new sub domain. Have read a few articles but having no luck editing the .htaccess as it might be interfering with all the rules in there. Old: www.example.co.uk New: https://secure.example.co.uk The current rules are quite handy because it seems to have sorted out the structure for all internal links. It has even updated the file path for images (or this could just be wordpress as the url was updated under general settings). This is the current .htaccess <files wp-config.php> order allow,deny deny from all </files> # BEGIN WordPress <IfModule mod_rewrite.c> RewriteEngine On RewriteBase / RewriteRule ^index\.php$ - [L] RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteRule . /index.php [L] </IfModule> # END WordPress

    Read the article

  • Restart Trouble Ubuntu 12.10

    - by Cristi
    I`ve installed Ubuntu 12.10 on 3 different machines: Acer AspireOne Netbook Lenovo T60 Notebook Desktop - P5P77 - Asus mainboard - I3 - porcessor I have the same issue on all of themm they won`t restart (Shut Down work), I can see the proceses are killed but it freezez on black screen total iresponsive. Even when I access the bios config to modify something then Save&Exit then restart I get the same thing. This is the same for all 3 machines. I also cleaned up the desktop system, last version of bios update, full HDD format, fresh win7 installed (practicly no trace of the old Ubuntu 12.10) still the same problem. Some time ago I had Ubuntu 11.10 on the desktop system, it worked perfectly. I also mention that I tried all combinations of ACPI settings in BIOS, unfortunately with no results. Please advice, I am a little desperate :)) Best regards, Cristian, Romania

    Read the article

  • Compiz setting migration from 12.04 to 12.10

    - by Maksim
    I just make migration my ubuntu 12.04 home folder to fresh installed ubuntu 12.10. And again have problem with compiz. It have no one my custom settings! How it's possible then full copy home dir not transfer settings? Those things began the compiz start play with digits on config folders. like .compiz-1 or .compiz. I have now idea then I have to use this -1 then not. But I've rename my compiz folders in .gconf and .compiz. Which not help. How get my setting back. And why this mess happening?

    Read the article

  • When JDeveloper IDE doesn't render the visual editor

    - by Frank Nimphius
    Though with Oracle JDeveloper 11g the problem of the IDE not rendering JSF pages properly in the visual editor has become rare, there always is a way for the creative to break IDE functionality. A possible reason for the visual editor in JDeveloper to break is a failed dependency reference, which often is in a custom JSF PhaseListener configured in the faces-config.xml file. To avoid this from happening, surround the code in your PhaseListener class with the following statement (for example in the afterPhase method) public void afterPhase(PhaseEvent phaseEvent) {   if(!ADFContext.getCurrent().isDesigntime()){ ... listener code here ... } } The reason why the visual editor in Oracle JDeveloper fails rendering the WYSIWYG view has to do with how the live preview is created. To produce the visual display of a view, JDeveloper actually runs the ADF Faces view in JSF, which then also invokes defined PhaseListeners. With the code above, you check whether the PhaseListener code is executed at runtime or design time.If it is executed in design time, you ignore all calls to external resources that are not available at design time.

    Read the article

  • Fail to load Ubuntu11.10 onto ASUS PC 1015PX

    - by strugglingbadly
    My new ASUS 1015PX has the usual windows on the SDA1 partition - 100Gb, followed by SDA2 - 15Gb - with the recovery for windows, SDA3 with 183Gb for drive 'D' for windows. ASUS uses SDA4 - 19Mb for it's own use. 11.10 on an USB will load Ubuntu on a try basis and it seems OK, but every time I attempt an install, Ubuntu quickly shuts down and the restart process begins. Repeat endlessly. I've tried 10.4 netbook but that will not boot at all with the machine reporting - unknown keyword in config file gfxboot vesamenu.c32: not a COM32R image. I've gone through the above about 6 times each with the same results. I have also tried the above with the partitioning unchanged, and using gparted, with SDA3 formatted to ext4, windows 'D' reduced to 8Gb providing a 175Gb unallocated space - all to no avail. Please help

    Read the article

  • Ubuntu 11.10 Desktop Sharing resolution

    - by Zuma123
    I have 11.10 running on a Netbook with Unity2D @ 1024x600 resolution. I VNC into my Ubuntu Netbook from Win7 @ 1920x1200. Is there a way make Desktop Sharing on 11.10 run at a resolution higher than the hardware resolution? I am not too worried about speed/lag. 1440x900 would be perfect. I like to keep my machine as clean as possible, I could add eg. vnc4server but I'd prefer not to do that. I have looked for but cannot find an appropriate vino-server config file. Thank you.

    Read the article

  • DesktopCouch does not sync

    - by burli
    I have here two computers, both installed with Ubuntu 10.10. The Laptop uses the Netbook Edition. Both are connected, file sync and Tomboy Notes are working. My Desktop syncs databases with Ubuntu One, my Laptop does not. If I delete for example the contacts Database on my Desktop, the database is back a view seconds later. If I delete the same database on my laptop, nothing happens. (I'm using Futon) If I kill beam, delete the config and try to restart with the dbus-send I got an error Error org.freedesktop.DBus.Error.NoReply: Did not receive a reply. Possible causes include: the remote application did not send a reply, the message bus security policy blocked the reply, the reply timeout expired, or the network connection was broken. How can I fix that?

    Read the article

  • Security settings for this service require 'Basic' Authentication

    - by Jake Rutherford
    Had an issue calling WCF service today. The following exception was being thrown when service was called:WebHost failed to process a request. Sender Information: System.ServiceModel.ServiceHostingEnvironment+HostingManager/35320229 Exception: System.ServiceModel.ServiceActivationException: The service '/InteliChartVendorCommunication/VendorService.svc' cannot be activated due to an exception during compilation.  The exception message is: Security settings for this service require 'Basic' Authentication but it is not enabled for the IIS application that hosts this service..Ensured Basic authentication was indeed enabled in IIS before getting stumped on what actual issue could be. Turns out it was CustomErrors setting. Value was set to "off" vs "Off". Would have expected different exception from .NET (i.e. web.config parse exception) but it works now either way.

    Read the article

< Previous Page | 205 206 207 208 209 210 211 212 213 214 215 216  | Next Page >