Search Results

Search found 41882 results on 1676 pages for 'png files'.

Page 9/1676 | < Previous Page | 5 6 7 8 9 10 11 12 13 14 15 16  | Next Page >

  • UIImage couldn't load some png picture.

    - by www.ruu.cc
    I Use the following code to load png image: UIImage *imageBack1 = [UIImage imageNamed:@"Bar1.png"]; UIImage *imageBack2 = [UIImage imageNamed:@"Bar2.png"]; imageBack1 work right when imageBack2's value is nil, Bar1.png and Bar2.png are located at the same place,but why Bar2.png couldn't be load?

    Read the article

  • Can I programatically determine if a PNG is animated?

    - by alex
    I have PNG (as well as JPEG) images uploaded to my site. They should be static (i.e. one frame). There is such thing as APNG. (it will be animated in Firefox). According to the Wikipedia article... APNG hides the subsequent frames in PNG ancillary chunks in such a way that APNG-unaware applications would ignore them, but there are otherwise no changes to the format to allow software to distinguish between animated and non-animated images. Does this mean it is impossible to determine if a PNG is animated with code? If it is possible, can you please point me in the right direction PHP wise (GD, ImageMagick)?

    Read the article

  • How to save optimized png images with java's ImageIO?

    - by Christoph
    I am generating lots of images in java and saving them through the ImageIO.write method like this: final BufferedImage img = createSomeImage(); ImageIO.write( img, "png", new File( "/some/file.png" ); I was happy with the results until Google's firefox addon 'Page Speed' told me that i can save up to 60% of the size if i optimize the images. The images are QR codes, their size is around 900B each and the firefox-plugin optimized versions are around 300B. I'd like to save such optimized 300B Images directly from java. So here my question again: How to save optimized png images with java's ImageIO?

    Read the article

  • Is PNG the most economically sound file format to store pictures in?

    - by raoulsson
    I am looking for an economically sound solution to store pictures long time. I read about the PNG file format that it has superior characteristics compared to JPEG, namely in these categories: no patents, no licenses, no royalities no quality loss yet compressed I have a lot of big ESP's from PhotoShop that contain tons of metadata, like layers and color profiles that I don't need to store (those were handy for the designer, when he worked with it). I want to convert these images without that hidden data, to a new target file format. Another side condition to my question is that the target file format has to be displayable in the browser. So I guess my options are limited anyway: GIF, JPEG, PNG. Am I missing something or is PNG the best fit for my case?

    Read the article

  • Prefix files with the current directory name using Powershell

    - by XST
    I have folders with images (*.png and *.jpg) >C:\Directory\Folder1 01.png 02.png 03.jpg 04.jpg 05.png And I want to rename all the files like this using powershell: >C:\Directory\Folder1 Folder1 - 01.png Folder1 - 02.png Folder1 - 03.jpg Folder1 - 04.jpg Folder1 - 05.png So I came up with this simple line: Get-ChildItem | Where-Object { $_.Extension -eq ".jpg" -or $_.Extension -eq ".png"} | rename-item -newname {$_.Directory.Name +" - " + $_.Name} If I have 35 or less files in the folder, I will have the wanted result, but if there is 36 or more files, I will end up with this: >C:\Directory\Folder1 Folder1Folder1Folder1 - 01.png Folder1Folder1Folder1 - 02.png Folder1Folder1Folder1 - 03.jpg Folder1Folder1Folder1 - 04.jpg Folder1Folder1Folder1 - 05.png The loop stops when the file's name exceeds 248 characters. Any ideas why it's looping?

    Read the article

  • How do I increase the open files limit for a non-root user?

    - by iCode
    This is happening on Ubuntu Release 12.04 (precise) 64-bit Kernel Linux 3.2.0-25-virtual I'm trying to increase the number of open files allowed for a user. This is for an my ecplise java application where the current limit of 1024 is not enough. According to the posts I've found so far, I should be able to put lines into /etc/security/limits.conf like this; soft nofile 4096 hard nofile 4096 to increase the number of open files allowed for all users. But, that's not working for me, and I think the problem is not related to that file. For all users, the default limit is 1024, regardless of what is in /etc/security/limits.conf (I have been rebooting after changing that file) $ ulimit -n 1024 Now, despite the entries in /etc/security/limits.conf I can't increase that; $ ulimit -n 2048 -bash: ulimit: open files: cannot modify limit: Operation not permitted The weird part is that I can change the limit downwards, but can't change it upwards - even to go back to a number which is below the original limit; $ ulimit -n 800 $ ulimit -n 800 $ ulimit -n 900 -bash: ulimit: open files: cannot modify limit: Operation not permitted As root, I can change that limit to whatever I want, up or down. It doesn't even seem to care about the supposedly system-wide limit in /proc/sys/fs/file-max # cat /proc/sys/fs/file-max 188897 # ulimit -n 188898 # ulimit -n 188898 So far, I haven't found any way to increase the open files limit for a non-root user, and I really don't want to be running my application as root. How should I properly do this? I have looked at all the posted and tried the given options but no luck!

    Read the article

  • Installer script removed a buch of 64-bit files and replaced them with 32-bit ones, how do I get them back?

    - by ILikePizza555
    I ran this installer script to get some drivers for my printer, however I noticed that it started removing some of my programs, as well as 64-bit system files, and started replacing them with 32-bit ones! Luckily, I managed to cancel the operation before it did any serious damage. I also noticed that offical PPA's for 32-bit files were added. Should I remove them? So how do I get these files back?

    Read the article

  • Only show hidden files in certain directories

    - by Joseph Silvashy
    On a Unix system (or more specifically on OS X) is it possible to show hidden files in only some directories? For example as a developer I want to see the hidden files in my Rails projects but not on my desktop as well. I guess I'm just tired of seeing all these .DS_Store and .trashes files swimming around, any remedies not directly related are welcome too!

    Read the article

  • Transfer files between VPC and Host Machine.

    - by gwc
    I have a virtual machine setup and I am able to transfer files from the host to the vpc and from the vpc to the host. I then copied the vpc to another machine and I am not able to copy files anymore. How can you setup the ability to copy files between the VPC and the host machine.

    Read the article

  • QNAP ftp: hide Mac OSX hidden files .DS_Store and others

    - by tobia.zanarella
    I am using a QNAP as an FTP server. Users access this NAS from their Mac OSX Lion clients and put files and folders inside the QNAP shares, which are later being accessed via FTP from external points. The problem is that inside the folders uploaded to the QNAP folders Mac OS leaves its .DS_Store and .* hidden files. I know that for a Samba share, for example, I need to add two lines to the /etc/samba/smb.conf file: veto files = /._*/.DS_Store/ delete veto files = yes But what can I do on a QNAP FTP service? Thank you.

    Read the article

  • Making folder available offline - Access denied (Windows XP)

    - by alex
    I have a drive on a server called "filestore" This has a folder called "users" in here, theres is a "users" folder. In there are folders such as - jsmith, jblogs, abrown - which correspond to usernames of our active directory. When a user logs on, the H: drive is mapped to their above mentioned folder - for example //filestore/users/jsmith I am trying to make one of these folders available offline in Windows XP. when i do so, i get an access is denied error. according to this - http://support.microsoft.com/kb/275461 i need to make the root folder have read access to the users. This means, however, all the users will be able to get in all the folders (if they browse to //filestore/users/) How can I still achieve offline files for that particular users "h:" drive, and still have the other folders secured?

    Read the article

  • SPS 2010 Prerequisite installer keep downloaded files

    - by elhombre
    Hi all In SharePoint 2010 there is a prerequisite installer. There are options to install the prerequisites from a local source. That's great, but is it also possible to tell the prerequitiinstaller to download the files and keep them in its prerequisite directory? I don't want to download the files manually as it is a big pain in the back to get these files from the Microsoft website.

    Read the article

  • VERY large text files and Snow Leopard

    - by cbmeeks
    Sometimes I need to work on EXTREMELY large text files. 200-300 megs or more. My favorite text editor on my MacBook Pro is TextMate. However, TM chokes on very large text files. Even ones around the 100MB mark. Is there a text editor that can handle such files for Snow Leopard? Thanks!

    Read the article

  • Automatically delete old files from a specified folder?

    - by davr
    I want to set up an automatic task that will automatically delete files older than X days from a specific folder. I'd prefer to do it without having to install any 3rd party software, but I'll be ok if it's a very small and simple utility. I want this to force me to keep my files organized. I am setting my browser to download to a standard folder, and then with this script to automatically delete files older than a week, I will be forced to move & organize downloaded files I want to keep.

    Read the article

  • Disable Offline Files (mobsync.exe) on Windows 7 Home

    - by Synetech
    This morning I was watching the CPU graph of a Windows 7 Home laptop and noticed that every few seconds, the CPU would spike several percent. I watched the processes and determined that it was mobsync.exe (Offline Files) that was the culprit. I tried the usual steps that Googling turns up, and clicking the Manage Offline Files link to bring up the Offline Files dialog to click Disable Synch does not work because the dialog will not display. This makes sense since everything I have read indicates that Offline Files is not even included/supported in the Home version, so I am at a loss as to why it is running at all, let alone why it is sucking up CPU cycles. (My best guess is that it was started when they pressed Win+X to access the Mobility Center.) Of course I can just kill mobsync, but it could always just come back. How/why would mobsync be running on a Home version and how can it be disabled (of course the Group Policy editor is not available on a Home version).

    Read the article

  • Batch renaming 32000 files - splitting into multiple subdirectories

    - by Gareth
    I've got a web server which has files uploaded to it. There is a script which assigns them numeric IDs and stores them in a corresponding subdirectory. I've now got 32000 of these uploads and that's too many for the server to handle in one directory. The script I'm using does have a way to "namespace" uploads so that ID 12345 - instead of sitting in /files/12345 - would sit in /files/namespaced/000/012/345. The code can deal with this just fine, but I now have 32000 subdirectories in the wrong naming style. What's the easiest way to go through my existing files and put them in the right place?

    Read the article

  • OAS log files filling up hard drive

    - by Andrew Hampton
    We've had issues with log files for Oracle Application Server filling up the hard drive on our server. The files are in the /network/admin folder and are named server.log_XXXXX.trc and client.log_XXXXX.trc where XXXXX are 5 digits. The files are typically anywhere from 1-2MB in size but can be up to 100MB and thousands of them are created at a rate of about 5-10 per minute. Does anyone know how to disable these logs? Thanks!

    Read the article

  • Windows Server 2008 R2 Finding Hidden files that are causing drive to fill up

    - by Xaxum
    What is the best way to find out what files are sucking up so much space? My Windows directory lists its size as 15 GB but when I run a powershell script against it, it shows 40 GB. It seems 40 GB is about right because if I use that 40 GB and add it to the rest of my folders on C I get the 45 GB that is in use. I checked volume shadow copy and that is disabled and there are no volume shadow files listed. My page file is limited to 3 GB. How can I find out what files are taking 25 GB of space? Final Results: We ended up disabling UAC and then the Properties showed the true file size of the Windows folder, as shown in powershell script. This allowed us to drill down to the folder that had the issues... In our case it was in C:\Windows\System32\config\systemprofile\AppData\Local\VMware\VDM there were several large dump files, these were the mysterious unknown files in WinDirStat. It turns out our nessus scan is crashing a view service for a couple seconds so it creates a large dump file. So we will try this.. https://discussions.nessus.org/thread/5212

    Read the article

  • View rotated log files Mac OS X Server (*.?.gz)

    - by Meltemi
    Trying to look at some of our older log files and find they're cryptic "Unix Executable Files". This particular server I'm working with is an older Mac OS X Server (10.4 - Tiger). -rw-r----- 1 root admin 36 1 Jun 15:48 wtmp -rw-r--r-- 1 root admin 578 27 May 17:40 wtmp.0.gz -rw-r----- 1 root admin 89 26 Apr 13:57 wtmp.1.gz -rw-r----- 1 root admin 78 29 Mar 16:43 wtmp.2.gz -rw-r----- 1 root admin 69 15 Feb 17:21 wtmp.3.gz -rw-r----- 1 root admin 137 16 Jan 13:09 wtmp.4.gz i'm using zless to try and view the contents of the .gz files. and what i see is unreadable: ... <DF>^R<AF>ttyp1^@^@^@joe54^@^@^@^@^@108.184.63.22^@^@^@^@K<DF>"<B8>ttyp1^@^@^@^@^@ ^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@K<DF>%<A1>console^@^@^@^@^@^@^@^@^@ ^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@^@K<E0>1 ~^@^@^@^@^@^@^@shutdown^@^@^@^@^@^@^@^@ ^@^@^@^@^@^@^@^@K<E0>1^L~^@^@^@^@^@^@^@reboot^@^@^@^@^@^@ ... same goes for system.log.0.gz, etc... anything that's been rolled in compressed .gz files. What am i missing?

    Read the article

  • Cygwin creates files with special (shared) icons on windows

    - by barjonah
    I use cygwin to transfer files between linux and windows machines. Everytime I transfer a file to a windows machine it adds an extra shared user icon on the file's or folder's icon itself. This also happens if I create a file from cygwin on windows using pretty much any command: echo, vim, nano, cat. This is what the cygwin-created folders (and files) look like. This is what a normal folder (or file) looks like. I'm thinking it has to do with permission, because I'd have to chmod it everytime if I want other applications to access the files or folders on windows. How can I tell cygwin to create regular ol' files just like a user or any program would?

    Read the article

  • best tool for searching within unstructured log files [closed]

    - by Alex Holding
    i am supporting a number of bespoke applications at the minute and searching through their very non standard logs is a nightmare, so im looking for a tool which can do the following - load large text files search through multiple files at once and display all results can search with regex can be used to view and search unstructured text files There are some great looking log tools available but they all seem to be focused on structured logs, where the logs i deal with most days are just flat text files. I am currently using notepad++ but that has its own annoyances so im hoping there is a dedicated log analysis tool that i havent found yet.

    Read the article

  • DirectX11 CreateWICTextureFromMemory Using PNG

    - by seethru
    I've currently got textures loading using CreateWICTextureFromFile however I'd like a little more control over it, and I'd like to store images in their byte form in a resource loader. Below is just two sets of test code that return two separate results and I'm looking for any insight into a possible solution. ID3D11ShaderResourceView* srv; std::basic_ifstream<unsigned char> file("image.png", std::ios::binary); file.seekg(0,std::ios::end); int length = file.tellg(); file.seekg(0,std::ios::beg); unsigned char* buffer = new unsigned char[length]; file.read(&buffer[0],length); file.close(); HRESULT hr; hr = DirectX::CreateWICTextureFromMemory(_D3D->GetDevice(), _D3D->GetDeviceContext(), &buffer[0], sizeof(buffer), nullptr, &srv, NULL); As a return for the above code I get Component not found. std::ifstream file; ID3D11ShaderResourceView* srv; file.open("../Assets/Textures/osg.png", std::ios::binary); file.seekg(0,std::ios::end); int length = file.tellg(); file.seekg(0,std::ios::beg); std::vector<char> buffer(length); file.read(&buffer[0],length); file.close(); HRESULT hr; hr = DirectX::CreateWICTextureFromMemory(_D3D->GetDevice(), _D3D->GetDeviceContext(), (const uint8_t*)&buffer[0], sizeof(buffer), nullptr, &srv, NULL); The above code returns that the image format is unknown. I'm clearly doing something wrong here, any help is greatly appreciated. Tried finding anything even similar on stackoverflow, and google to no avail.

    Read the article

  • blurry images with mogrify & convert

    - by user140393
    Does anyone know why this image is so blurry? I did a convert from pdf to png and it turned out like that. Before deleting imagemagick and it's entire toolset from software-center most of my image programs were displaying like this image. Now for the most part it's just blurry, though a couple still display like that such as gimp. I am running in xfce maybe it's to do with the distribution enviornment. Main issue is the absurd blurriness. I reinstalled all additional packages that were available for imagemagick in the software-center I use convert *.pdf *.png & mogrify -format png *.pdf to convert Now on the other hand if I converted the file to djvu and converted that to a tif. The images have no problem converting. More so it does not generate an oversized tif file of around 25mb compared to 3mb with djvu which is super clear & no blurriness.

    Read the article

< Previous Page | 5 6 7 8 9 10 11 12 13 14 15 16  | Next Page >