Search Results

Search found 40479 results on 1620 pages for 'binary files'.

Page 410/1620 | < Previous Page | 406 407 408 409 410 411 412 413 414 415 416 417  | Next Page >

  • brackets in php templates

    - by Othman
    I have a script, and now it's the perfect time to have an option to change theme templates. I was looking forward to store these files in php files, or html files. with CSS file. I looked over some codes and I found that they use brackets to refer to the variable. Here's an example of a template block: <div class='block_div'> <center>{title}</center> {content} </div> I know that I can use PHP vars usering <?=$var;?>. However, I think the one above is better, and I looked over the web to find anything about it but I couldn't. My Question is how I can use these brackets in my php or html code. to let the user changing the template only not the code!

    Read the article

  • php.ini modifications has no effect on joomla

    - by user488736
    I have a netbook with a recently installed ubuntu server 12.04 64-bit. I have (only) installed Joomla on it (version 2.5) and now I am trying to tweak the php.ini settings so I can install jomsocial onto joomla. I have changed these settings: upload_max_filesize = 2M post_max_size = 8M to this: upload_max_filesize = 20M post_max_size = 20M in all php.ini-files in my system. Thats six files, found using 'locate': /etc/php5/apache2/php.ini /etc/php5/cli/php.ini /usr/share/doc/php5-common/examples/php.ini-development /usr/share/php5/php.ini-production /usr/share/php5/php.ini-production-dist /usr/share/php5/php.ini-production.cli I have double-checked that these files actually contain the new settings. Still, joomla does not recognize the new settings, stating that upload_max_filesize is still only 2M and post_max_size is still only 8M The same is true after a complete restart of the server. What am I doing wrong here?

    Read the article

  • On-the-fly file search in gnome 3?

    - by jan groth
    I recently switched to Gnome 3 (on Onereic 11.10). I was a big fan of gnome-do (on Gnome 2), where I heavily used the "files and folders"-plugin. This allows to configure a list of files and folder to be searched on-the-fly while typing. E.g. starting to type "timesheet" opens "timesheet_2011.ods" from my dropbox account... I can't find anything like this in Gnome 3. I know that there is a list of "recent items", and also the name of bookmarked folders in nautilus is searched, but that's not (exactly) what I'm looking for... So, what I want is this: tell Gnome 3 where my files are (certainly not only in my home folder) have them searched on the fly while I'm typing I think Unity offers something like this - but hey, I'm using Gnome 3 ;-) Anything I missed?

    Read the article

  • Strategies for removing register_globals from a file

    - by Jonathan Rich
    I have a file (or rather, a list of about 100 files) in my website's repository that is still requiring the use of register_globals and other nastiness (like custom error reporting, etc) because the code is so bad, throws notices, and is 100% procedural with few subroutines. We want to move to PHP 5.4 (and eventually 5.5) this year, but can't until we can port these files over, clean them up, etc. The average file length is about 1000 lines. I've already cleaned up a few of the low-hanging fruit, however the job took almost an entire day for 2 300-500 line files. I am in a quagmire here (giggity). Anyway, has anyone else dealt with this in the past? Are there any strategies besides tracing backwards through the code? Most static analysis tools don't look at code outside of functions - are there any that will look at the procedural code and help find at least some of the problems?

    Read the article

  • chmod 700 and htaccess deny from all enough?

    - by John Jenkins
    I would like to protect a public directory from public view. None of the files will ever be viewed online. I chmoded the directory to 700 and created an htaccess file that has "deny from all" inside it. Is this enough security or can a hacker still gain access to the files? I know some people will say that hackers can get into anything, but I just want to make sure that there isn't anything else I can do to make it harder to hack. Reply: I am asking if chmod 700 and deny from all is enough security alone to prevent hackers from getting my files. Thanks.

    Read the article

  • Efficient way to check for changes to the contents of folders

    - by MrVimes
    I am creating an application that maintains a database of files of a certain type in a given folder (and all subfolders) Initially the program will recurse the folders and add any file it finds of that type to the database. I want the application to have the ability to re-scan the folder and add any files that were not there the last time the folders were scanned. It can't use the date created property of the file because there is a high chance of a file being added to the folders that isn't a new file. I am wondering what the most efficient way of doing this is, and if there is a way that doesn't involve checking each file is in the database already (which, if there are 5000 files would mean 5000 queries of a list 5000 items in size, or 25 million 'checks' for the sql engine to perform) I suppose a more specific question to acheive the same goal would be - is there a property of a file (in Microsoft Windows) that will reliably tell you when that file arrived in that folder.

    Read the article

  • Whats a good structure to save and retrieve locations of images?

    - by Goot
    I got a java-ee application, where I collect informations about movies. Im my backend I provide data like the name, description, genre and a random uuid. I also got lots of related files, which are stored on a file server. Including some screenshots, the dvd or bluRay cover and video trailers. My current approach is: When saving the files to the fileserver, I retrieve the movies random uuid (which is the primary key btw.). I then rename the files screenshot_[UUID]_1, screenshot_[UUID]_2 ... etc. Now, there are lots of other ways to handle this, like saving all filenames in a database or creating a dir structure on the fileserver for every uuid and, e.g., return all images in the "[uuid]/screenshots" folder via REST. I expect about 30k requests a day, so the service has to be pretty performant. Whats the best way to solve this?

    Read the article

  • At what point is asynchronous reading of disk I/O more efficient than synchronous?

    - by blesh
    Assuming there is some bit of code that reads files for multiple consumers, and the files are of any arbitrary size: At what size does it become more efficient to read the file asynchronously? Or to put it another way, how small must a file be for it to be faster just to read it synchronously? I've noticed (and perhaps I'm incorrect) that when reading very small files, it takes longer to read them asynchronously than synchronously (in particular with .NET). I'm assuming this has to do with set up time for things like I/O Completion Ports, threads, etc. Is there any rule of thumb to help out here? Or is it dependent on the system and the environment?

    Read the article

  • The Best How-To Geek Articles for November 2012

    - by Asian Angel
    Last month we covered topics such as why 64-bit Windows needs a separate “Program Files (x86)” folder, how to uninstall your Windows product key before selling your PC, how to deal with locked files in Windows, and more. Join us as we look back at the best articles for November. HTG Explains: Does Your Android Phone Need an Antivirus? How To Use USB Drives With the Nexus 7 and Other Android Devices Why Does 64-Bit Windows Need a Separate “Program Files (x86)” Folder?

    Read the article

  • Is Azure Compatible with JPEG XR?

    - by Shawn Eary
    I just put an F#/MVC app into a Windows Azure solution as a Web Role. Before migration, my JPEG XR (*.WDP) files were getting displayed on the client in IE9 without issue via my local and hosted sites. Now, after migration into Windows Azure, my JPEG XR files neither get displayed in my local Windows Azure compute emulator nor do they get displayed when they are deployed to http://*.cloudapp.net. Is there some sort of conflict with Widows Azure and (JPEG XR) *.wdp files? If so, what is the accepted best practice for overcoming this conflict?

    Read the article

  • How to make Ubuntu LiveCD be able to use USB Flash drive and external hard drive?

    - by ????
    I am booting up Ubuntu 2012.04 LiveCD... and was able to do /sudo mount /dev/sda1 /mnt and be able to see files in /mnt, which is the main hard drive that can't boot up any more. So to copy files from /mnt to an external hard drive or USB flash drive, I connected a 1TB external hard drive and 2 USB flash drives to the computer, but for some reason, in "File Systems", I can't drag and drop files from /mnt into those external hard drive or USB flash drives? I can't open or look into those drives either... How to make it work?

    Read the article

  • I loose some directories when i upgrade from Ubuntu 11.10 to 12.04

    - by maythux
    last day i upgraded my ubunut 11.10 desktop to ubuntu 12.04. I was running a KVM virtual about 7 machines and managed by virt-manage software.... anyway when i finished upgrading i found that virt-manager is not working so i have to reconfigure it again and install some other missing packages that was deleted!!!! anyway i solve this issue...then i started to restore my virtual machines i restore 2 machines without any problems the third and fourth ones (windows) make a check disk that takes more that 6 hours but finally it works... other machines i cant find their attached hard disks i don't know what happens but i cant found that files. 1- upgrading delete files!?!! 2- Is there anyway to restore those files? thanks in advance

    Read the article

  • Restoring Virtualbox machine images from old hard drive

    - by memilanuk
    I recently replaced the HDD in my laptop, and re-installed Windows & Ubuntu. Now I want to restore the various virtual machines I had set up on the old HDD, which is mounted in an external USB enclosure. I can read the HDD okay, and the 'bad' spots seemed to be in the Windows partition... but whenever I try to restore the VDI files the copy errors out. I've tried drag-n-drop in Nautilus, I've tried grsync, etc. Always bombs out on the VDI files. I've copied over multi-GB dvd iso images with no problem, but the VDI files always fail the checksums. Any ideas? TIA, Monte

    Read the article

  • Dynamic content realoding

    - by Kikaimaru
    Is there a relatively simple way to dynamicaly reload content files? (ie: effect files) I know i can do following: Detect change of file Run content pipeline to rebuild that specific file Unload ALL content that was loaded Load All content And use double references to reference content files. Problem is with step 3 (and step 2 isn't that nice too). But i need to unload everything because if i have model Hero.x which references Model.fx effect, and i change Model.fx file, i need to reload Hero.x file which will then call LoadExternalReference on Model.fx. So I guess question is, did someone mange to make this work without rewriting whole ContentManager (and every ContentReader) and tracking calls to LoadExternalReference?

    Read the article

  • How can set up a file sharing server to allow remote access over a WAN?

    - by pst007x
    I want to set up a server to allow remote access over a WAN. The purpose is to allow people to remote connect and download specific files from a specified folder hosted on the server. Is there software available to achieve this, preferably with an interface so I can easily add and remove users, plus monitor who is connected at any time? I have looked around and only found terminal references, and not very clear how to set up. I have a service running on my Win server, but I am hoping to move from that to Ubuntu for my file sharing. I do not want people who access the server to have any local accounts or any access to any folder/files apart from authorised folders/files. Thanks

    Read the article

  • Which Hosting for a Mobile Infotainment Portal? [closed]

    - by VenomVipes
    Possible Duplicate: How to find web hosting that meets my requirements? I am building up a portal for music and movie downloads. My target is 85% Mobile User. I expect (acc to our promotion plan) 1000 Visitors a day and at least 300 parallel downloads. Files are mp3 & 3gp format. Please suggest the type of hosting Cloud/VPS/Dedi. Please suggest config (suppose I have 50Gb Video & 30Gb Audio). We have a budget of around 70$/m for Server. Also suggest if I should keep all files in my HostServer or just put links to direct downlad frm other sites offering the same video. My goal is Fast Loading on Mobile Browser (2G GPRS). Also faster download of files.

    Read the article

  • Errors with linux-image-3.8.0-36-generic and GRUB

    - by user285239
    OS: Ubuntu 12.04 LTS Problem 1: When installing anything (or updating) it always ends with this error: Errors were encountered while processing: linux-image-3.8.0-36-generic linux-image-3.8.0-38-generic linux-image-generic-lts-raring linux-generic-lts-raring Problem 2: When installing or updating stuff it opens some grub files and halts execution until I close these grub files. See screenshot below. Grub files + terminal window (screenshot) Pastebin with terminal output while trying to install something I should mention that I don't known if these two errors are related, but it is a fact that every time I try to install something or run an update both of the above errors pop up. I couldn't find anything of interest in the logs I looked at (probably because I don't know what/where to look), but tell me if you need me to upload something. Edit 02 june. This is the output from lsb_release -a kasper@ubuntuRW:~$ lsb_release -a No LSB modules are available. Distributor ID: Ubuntu Description: Ubuntu 12.04.4 LTS Release: 12.04 Codename: precise

    Read the article

  • apt-get not recognizing downloaded archives

    - by meteors
    I installed Ubuntu Gnome 13.10. I previously had Ubuntu Gnome 13.04 and had all the archives in the /var/cache/apt/archives/ stored to a removable disk. After installing 13.10 I copied all my archives to the above mentioned path. When I run apt-get install it tries to fetch the archives although I have the archives. Also if instead of apt-get install if I try to install individual .deb files using dpkg -i everything runs fine. These are the permissions of files: How do I fix this. Previously copying archives like this worked fine and downloading duplicates the files.

    Read the article

  • How do I save to an NTFS partition?

    - by RADHAKRISHNAN
    I was using ubuntu 11.04 on my Laptop. While installing it from a DVD, I have created a 10 GB NTFS partition at the beginning of the hard disk, as primary. All other partitions ( swap, a ext3, a ext4 and a FAT32 ) are created in as logical in the extended partition. All were working well in ubuntun11.04. Now the system was upgraded to ubuntu 11.10 via internet and was sucessful. But unable to either create folder/files or to write to existing files in the said NTFS partiton. But files in the partition can read - means mounting done. Same is the case even if logged in as root also. Fortunately no such problem with other partitions including FAT Why it is so, please help.

    Read the article

  • 12.04 LTS won't install from CD

    - by Rob Hays
    I've been trying to install 12.04 LTS onto a Dell with a PIII from CD. Booting from the CD the install gets through the "Who are you" process, begins copying files. The progress bar gets as far as the last period in "Copying files...". The box clears, and an error box comes up "The installer has encountered an unrecoverable error. A desktop session will now be run so that you may investigae the problem or try installing again." When I try to install from this desktop session, the install gets to the same point, the copying files box closes, and then just stops. The pointer is busy, the cd drive spins up occaisonally with no data transfer, no hard drive activity. When I boot from the CD and access the disk boot menu, the disk checks good, memory checks good ( I upgraded the original memory to 512 mb). I also updated the bios to the newest from Dell. This is an older L866r, but should meet the requirements.

    Read the article

  • Laptop won't get to grub after meddling with kernels

    - by Joren
    I recently found out I had different kernel versions installed, so I removed the 3.9 kernel as proposed in this answer on my earlier question. After that, I updated my grub, but My laptop won't seem to reach it. Output of aptitude search linux | grep ^i: i libselinux1 - SELinux runtime shared libraries i libselinux1:i386 - SELinux runtime shared libraries i linux-firmware - Firmware for Linux kernel drivers i A linux-headers-3.11.0-12 - Header files related to Linux kernel versi i A linux-headers-3.11.0-12-generic - Linux kernel headers for version 3.11.0 on i linux-headers-3.11.4-031104 - Header files related to Linux kernel versi i linux-headers-3.11.4-031104-gen - Linux kernel headers for version 3.11.4 on i linux-headers-generic - Generic Linux kernel headers i linux-libc-dev - Linux Kernel Headers for development i linux-sound-base - base package for ALSA and OSS sound system i pptp-linux - Point-to-Point Tunneling Protocol (PPTP) C i syslinux - collection of boot loaders i syslinux-common - collection of boot loaders (common files) i syslinux-legacy - Bootloader for Linux/i386 using MS-DOS flo i util-linux - Miscellaneous system utilities How do I proceed to fix this error from my live cd?

    Read the article

  • Accessing network shares through ASP.Net

    - by jkrebsbach
    In my impersonation enabled web site I needed to access files on a network share.  Running locally, everything worked fine. After deploying out to the dev server and hitting the web site from my PC, things fell apart. With impersonation enabled, we can access files on the server itself, but a network share is another story.  Accessing a share on another server, we encounter the infamous "double hop" situation, were the credentials have already been digested on the web server and are not available for the file server. We need to either expose the shared files to the identity IIS is running under, or create a new impersonation context.

    Read the article

  • How to organize functional programs

    - by bleakcabal
    In OOP, your basic unit of organization for code is the class. A frequently used methodology in Java, C# and similar languages is to organize your code around having one file for each class with the file name following the class name. You can consider each of these class as a unit of organization to group a single concept. These classes are in in namespaces which often follow the directory structure of the files in the solution/project. Namespaces are another level of organization. How are large projects in functional languages typically organized? How to you determine how to split your functions into different files? Are other units of grouping beside files used? How is code typically organized within a single file?

    Read the article

  • How to store and update data table on client side (iOS MMO)

    - by farseer2012
    Currently i'm developing an iOS MMO game with cocos2d-x, that game depends on many data tables(excel file) given by the designers. These tables contain data like how much gold/crystal will be cost when upgrade a building(barracks, laboratory etc..). We have about 10 tables, each have about 50 rows of data. My question is how to store those tables on client side and how to update them once they have been modified on server side? My opinion: use Sqlite to store data on client side, the server will parse the excel files and send the data to client with JSON format, then the client parse the JOSN string and save it to Sqlite file. Is there any better method? I find that some game stores csv files on client side, how do they update the files? Could server send a whole file directly to client?

    Read the article

  • Ubuntu will not start any more !!!!!!!! PLEASE HELP?

    - by mike
    I really need help , I left my computer downloading all the night and i did donwload 35 go of movies ( legal ....) , I restart the computed in the morning then i booted in my encrypted windows partition for my work. Surprise in the evening when i try to boot on linux again , then it doesnt start as it's telling me low graphic mode , and doesnt boot( happen when it's really full ) . Tryed in rescue and it's telling me i have 0 mo free. Tryed in shell comand sdo root rm - Some files IMPOSSIBLE . it's telling me that my files are only in read only file systems. I mouted my other hard drive in windows as well but there is a '' Write protection '' i can only read the files. Please let me know , what solution do i have ? Should i try live usb with ubuntu ? Thanks guys

    Read the article

< Previous Page | 406 407 408 409 410 411 412 413 414 415 416 417  | Next Page >