Search Results

Search found 41882 results on 1676 pages for 'png files'.

Page 264/1676 | < Previous Page | 260 261 262 263 264 265 266 267 268 269 270 271  | Next Page >

  • Disable "Ubuntu 14.04" and "Files" from auto sorting alphabetically and THEN<(Disable the then.) automatically going to that location

    - by Doodski Man
    I looked about in Ubuntu 14.04 and I could not find a option to disable a feature that made extra work for me. After renaming something "Files" then reorganizes alphanumerically and then automatically moves the prompt/cursor/highlighting to the new file name at the new alphanumerical location. It's much easier and practical if the working cursor remains in the same area where the file was found that needed name changes or needed name spelling corrections. Windows 7 does what I ask about and it's a huge improvement over that auto sort and then go to the new file location. A example to consider is a file is in the "A" area and then you rename it to a "Z" and you are then taken to the "Z" stuff and need to scroll hundreds of lines back to the "A" section. It's annoying. Thanks for reading and taking the time.

    Read the article

  • Which app can I use for easily annotating pictures/screenshots?

    - by koushik
    Often I need to annotate (draw some arrows, lines, basic shapes like squares, ellipses etc and enter some text) on top of pictures (JPG, PNG images) and screenshots (again png images). I would also need to be able to crop, resize etc. I tried the Gimp but I could only enter text and perform all image transformations but couldn't find a way to draw boxes etc. I finally settled to Openoffice.org draw, but I know that isn't what I want, because in oodraw I need to insert my pic into a drawing and resize it (or the drawing) to fit and then go about making changes and finally export to png... Is there any image editor that allows adding shapes and text to jpg & png files and save the modified file in its place? If the tool can also have template collections (like dia does) for shapes that is an added bonus.

    Read the article

  • How to execute a scipt just by double clicking like .EXE files in Windows?

    - by maythux
    How can I make a bash script executable by double clicking just like .exe files in Windows? I tried creating a launcher and assigning the script to it, but there are two consequences: the terminal twinkles, disappears, and nothing is done. you must specify to run in terminal in order to work. I have a script that installs tomcat on an offline PC, including all dependencies of tomcat in the script. I need to make the script work on double clicking like windows since most who use the script will not be familiar with Ubuntu. Forget the above explanation. I want to make a script that can be run by double-clicking on it, without using the terminal. Anybody knows how?

    Read the article

  • How to export changed files in dependency project when create updates? (TortoiseSVN + VisualSVN)

    - by Relax
    I have a project with it depends on another project Core, i use svn:external to maintain the dependency. when i go to create an update of my project, i export those changed files of two tags of my project, but the changed files in Core won't be exported. For example, my project 2.0 depends on Core 1.0, where 3.0 depends on Core 2.0, then i create an update, changed files of my project 3.0 will all be exported, but none of those in Core 2.0 This can be done thru a two-steps process, first export changed files of the project 3.0, then the Core 2.0, it is ok if i can remember it everytime when i create an update, but in case i forget someday, the broken update will most probably cause a crash. I wonder, is there a way to cover this situation in TortoiseSVN, will let me do it in one operation?

    Read the article

  • SVN - How to export changed files in dependency project when create updates?

    - by Relax
    I have a project with it depends on another project Core, i use svn:external to maintain the dependency. when i go to create an update of my project, i export those changed files of two tags of my project, but the changed files in Core won't be exported. For example, my project 2.0 depends on Core 1.0, where 3.0 depends on Core 2.0, then i create an update, changed files of my project 3.0 will all be exported, but none of those in Core 2.0 This can be done thru a two-steps process, first export changed files of the project 3.0, then the Core 2.0, it is ok if i can remember it everytime when i create an update, but in case i forget someday, the broken update will most probably cause a crash. I wonder, is there a way to cover this situation in TortoiseSVN, will let me do it in one operation? I use TortoiseSVN + VisualSVN

    Read the article

  • How to add JS/CSS files to Joomla modules?

    - by Apache Fan
    I am starting out with Joomla and am writing a simple Joomla module. I am using some custom CSS and JS in my module. Now when I distribute this module I need my JS/CSS files to go with the ZIP. I have added my files in my module ZIP file. This is what i need to know - How do I refer to these CSS/JS files in my module so that even if I distribute the module as a zip i would not have to send the css/js files separately? I tried looking at different solutions including http://www.howtojoomla.net/how-tos/development/how-to-add-cssjavascript-to-your-joomla-extension But I was not able to figure out what the URL for the JS/CSS file should be? I am using Joomla 1.7 hosted on a cloud hosting site. Thanks

    Read the article

  • how to get powershell to look for files in other folders when moving items?

    - by steeluser
    I have written this script to move files to the destination folder. Looks like I am missing something here because when I run the script, it is only looking for .zqx files in current directory and not all the drives. Please note that the ( dir $paths..) part is returning the list of .zqx files promptly. Paths.txt has drive letters like this C:\ D:\ E:\ $paths = get-content paths.txt mv (dir $paths -r -fi *.zqx | ?{$_.lastwritetime -lt ($sevendaysold)}) -dest e:\xqz Thanks Steeluser

    Read the article

  • Get root access for copying files to /usr/share/...?

    - by Vinaychalluru
    To be short, I want to copy a folder to a location /usr/share/screenlets/..... I don't know how to do it. I am using Ubuntu 10.04. I tried by logging in as root from terminal giving "su with my password". I even changed my user account type to ADMINISTRATOR by giving the root password when asked, yet, no use. Think all of you know that, even the option PASTE in the context menu's list in the folder "/usr/share/..." is INACTIVE. How can I copy those files?

    Read the article

  • Software to cd (change directories) into .jar/.ear files?

    - by Segphault
    Is there any software/script that will allow me to cd (change directories) into .jar/.ear/.zip files and edit the contents of the files it contains? I'm working on a large EJB project (yuck), and I frequently find myself in situations like the following: something.ear/ |-- something.jar/ | `-- fileINeedToEdit.xml I work primarily via the command line (Mac/Linux), so I find myself decompressing the files with jar -xvf, editing the file I need to edit, and then recompressing with jar -cvf. Obviously, this becomes a major headache after the first few times. I'd like to be able to treat the compressed files as directories, and simply cd (or some alternate command) to the file I want to edit. Does anyone know how I can accomplish this?

    Read the article

  • symlink files newer than X age, then later remove symlink once file ages?

    - by bleomycin
    Hello everyone, i have a large number of files/folders coming in each day that are being sorted automatically to a wide variety of folders. I'm looking for a way to automatically find these files/folders and create symlinks to them all within an "incoming" folder. Searching for file age should be sufficient for finding the files, however searching for age and owner would be ideal. Then once the files/folders being linked to reach a certain age, say 5 days, remove the symlinks to them automatically from the "incoming" folder. Is this possible to do with a simple shell or python script that can be run with cron? Thanks!

    Read the article

  • Silverlight, MSBuild, VS and some shared files. How?

    - by asgerhallas
    I have a VS project used for my .NET WCF host with some simple DTOs in it. I then have another project targeted for Silverlight with links to the files from the .NET-project. What's the best way automate the build, so that all files from the .NET project are automatically built to a Silverlight assembly too? I have tried the following in the Silverlight-library project: <Compile Include="..\KSLog.Core.Services.Shared\**\*.cs" Exclude="..\KSLog.Core.Services.Shared\Properties\AssemblyInfo.cs"></Compile> But when I do a build or a rebuild of the solution new files in the .NET project are not automatically added to the Silverlight project, and if I have deleted files in the .NET project, I get a compile error, saying the file is not found in the Silverlight project. Can I make it automatically update it self in some way? Or am I doing it all wrong?

    Read the article

  • What to do if I hate C++ header files?

    - by BlaXpirit
    I was always confused about header files. They are so strange: you include .h file which doesn't include .cpp but .cpp are somehow compiled too. NOTE: I UNDERSTAND EVERYTHING ABOUT THE HEADERS, PLEASE DON'T TELL ME I'M STUPID OR SHOULD USE OTHER LANGUAGE Recently I joined a team project, and of course, both .h and .cpp are used. I understand that this is very important, but I can't live with copy-pasting every function declaration in each of multiple classes we have. How do I handle the 2-file convention efficiently? Are there any tools to help with that, or automatically change one file that looks like example below to .h and .cpp? (specifically for MS VC++ 2010) class A { ... Type f(Type a,Type b) { //implementation here, not in another file! } ... }; Type f(Type a) { //implementation here } ...

    Read the article

  • Apache .shared folder

    - by Kevin
    There are already a bunch of rules in my Apache configuration. What I want to add is the following. There are some shared folders (.shared): /var/www/.shared/ and /var/www/.include/.shared/ and /var/www/.include/(.*)/.shared/. Now when someone visits http://domain.com/test.png it first executes the existing apache rules and will (when the file/folder was not found) look in those .shared folders. So suppose I've got this filesystem: /var/www/.shared/dog.png /var/www/.shared/test.gif and /var/www/domain.com/dog.png. Now when someone visits http://domain.com/test.gif, it must load the test.gif from the .shared folder. Now when someone visits http://domain.com/dog.png it must load the dog.png from the domain.com folder (because the existing apache rules will be executed first).

    Read the article

  • Which VCS is efficient for storing small changes to big text files?

    - by MiguelM
    A government agency publishes a text file with thousands of records. The entire file is about 60MB. Every day the file has about 60 new or changed records. We need to validate some info against that file, and for auditing purposes, we may need to recover the file as it was in certain date. I thought I could use a VCS to store the file using git, but I understand git stores the entire text file gzipped, so the 60BM file will still need about 30MB to store everyday version. I wonder if one of the free VCS stores only diff/patch files.

    Read the article

  • should I put the jar dependancies on the shared repo or just the source files ?

    - by ashy_32bit
    We have a Scala project with few source files (in Scala and Java) and quite some dependencies in various binary formats (jar and DLL). I'm wondering what should go into our shared git repo. Only the source files (developers have to download or somehow resolve the dependencies themselves) or the both the source files and the dependencies? I may add that dependencies are all third parties and available for download for free.

    Read the article

  • What are the best tools to help work with large ant files.

    - by klfox
    I just started working at a company that has a very large ant build file that imports lots of other large/small ant files. Needless to say it's giving me a headache trying to figure out what is going on. What are the best tools out there for: Getting some kind of concise answer on what is happening Visualizing the various targets Seeing performance on tasks Can be multiple tools. Any other tips/suggestions? I tagged this as java since I don't have the reputation to create an ant tag.

    Read the article

  • How to test obtaining a list of files within a directory using RSpec?

    - by John Topley
    I'm pretty new to the world of RSpec. I'm writing a RubyGem that deals with a list of files within a specified directory and any sub-directories. Specifically, it will use Find.find and append the files to an Array for later output. I'd like to write a spec to test this behaviour but don't really know where to start in terms of faking a directory of files and stubbing Find.find etc. This is what little I have so far: it "should return a list of files within the specified directory" do end Any help much appreciated!

    Read the article

  • Fourth texture = segmentation fault

    - by Robin92
    I keep on getting segmentation fault each time I load fourth texture - what type of texture, I mean filename, does not matter. I checked value of GL_TEXTURES_STACK_SIZE which turned out to be 10 so quite more than 4, isn't it? Here're code fragments: funciton to load texture from png static GLuint gl_loadTexture(const char filename[]) { static int iTexNum = 1; GLuint texture = 0; img_s *img = NULL; img = img_loadPNG(filename); if (img) { glGenTextures(iTexNum++, &texture); glBindTexture(GL_TEXTURE_2D, texture); glTexParameteri(GL_TEXTURE_2D,GL_TEXTURE_MIN_FILTER,GL_LINEAR); glTexParameteri(GL_TEXTURE_2D,GL_TEXTURE_MAG_FILTER,GL_LINEAR); glTexImage2D(GL_TEXTURE_2D, 0, img->iGlFormat, img->uiWidth, img->uiHeight, 0, img->iGlFormat, GL_UNSIGNED_BYTE, img->p_ubaData); img_free(img); //it may cause errors on windows } else printf("Error: loading texture '%s' failed!\n", filename); return texture; } actual loading static GLuint textures[4]; static void gl_init() { (...) //setting up OpenGL /* loading textures */ textures[0] = gl_loadTexture("images/background.png"); textures[1] = gl_loadTexture("images/spaceship.png"); textures[2] = gl_loadTexture("images/asteroid.png"); textures[3] = gl_loadTexture("images/asteroid2.png"); //this is causing SegFault no matter which file I load! } Any ideas? Problem is present on both Linux and Windows.

    Read the article

  • should I put the jar dependancies on the sahred repo or just the source files ?

    - by ashy_32bit
    We have a Scala project with few source files (in Scala and Java) and quite some dependencies in various binary formats (jar and DLL). I'm wondering what should go into our shared git repo. Only the source files (developers have to download or somehow resolve the dependencies themselves) or the both the source files and the dependencies? I may add that dependencies are all third parties and available for download for free.

    Read the article

  • Why isn't the pathspec magic :(exclude) excluding the files I specify from git log's output?

    - by Jubobs
    This is a follow-up to Ignore files in git log -p and is also related to Making 'git log' ignore changes for certain paths. I'm using Git 1.9.2. I'm trying to use the pathspec magic :(exclude) to specify that some patches should not be shown in the output of git log -p. However, patches that I want to exclude still show up in the output. Here is minimal working example that reproduces the situation: cd ~/Desktop mkdir test_exclude cd test_exclude git init mkdir testdir echo "my first cpp file" >testdir/test1.cpp echo "my first xml file" >testdir/test2.xml git add testdir/ git commit -m "added two test files" Now I want to show all patches in my history expect those corresponding to XML files in the testdir folder. Therefore, following VonC's answer, I run git log --patch -- . ":(exclude)testdir/*.xml" but the patch for my testdir/test2.xml file still shows up in the output: commit 37767da1ad4ad5a5c902dfa0c9b95351e8a3b0d9 Author: xxxxxxxxxxxxxxxxxxxxxxxxx Date: Mon Aug 18 12:23:56 2014 +0100 added two test files diff --git a/testdir/test1.cpp b/testdir/test1.cpp new file mode 100644 index 0000000..3a721aa --- /dev/null +++ b/testdir/test1.cpp @@ -0,0 +1 @@ +my first cpp file diff --git a/testdir/test2.xml b/testdir/test2.xml new file mode 100644 index 0000000..8b7ce86 --- /dev/null +++ b/testdir/test2.xml @@ -0,0 +1 @@ +my first xml file What am I doing wrong? What should I do to tell git log -p not to show the patch associated with all XML files in my testdir folder?

    Read the article

  • How much information do pdb files contain? (C# / .NET)

    - by Silver Dragon
    Is is wise to redistribute the pdb files along with a commercial application? Occasionally, I'm using the stack trace to get a more detailed error reporting logs from the deployed applications; can this functionality be achieved without relying to those files? Also, how much hints of the original source code does these files contain? Would it be easier to reverse-engineer my application using it?

    Read the article

  • Linux How to print all the files with the same prefix after searching for them?

    - by Alyx
    I need to search through a directory which contains many sub directories, each which contain files. The files read as follows question1234_01, where 1234 are random digits and the suffix _01 is the number of messages that contain the prefix, meaning they are apart of the same continuing thread. find . -name 'quest*' | cut -d_ -f1 | awk '{print $1}' | uniq -c | sort -n example output: 1 quest1234 10 quest1523 This searches for all the files then sorts them in order. What I want to do is print all the files which end up having the most occurrences, in my example the one with 10 matches. So it should only output quest1523_01 - 11

    Read the article

  • Why use binary files to stack up different versions on DMSs?

    - by edgarator
    I've used both Liferay and Alfresco trying to use them as the Document Management System for an intranet. I noticed the following: They use the file system and the database to store files They use a GUID to name the file on the filesystem and that GUID is used as an Id in the database. The GUID-named file is a binary file The GUID-named binary file stores all versions for a given file The path for the file in the DMS doesn't match the one in the file system The URL makes reference to the GUID when a certain file is requested What I want to know is why is this, and what would be the best way of doing it. Like how to would you create the binary file (zip?), and what parts would you keep in the binary file and what parts would you store in the database (meta-data, path?). I'm assuming some of the benefits of doing it like this. As having the same URL for a file, regardless of its current document path. And having only one file even if the file has changed names over time.

    Read the article

< Previous Page | 260 261 262 263 264 265 266 267 268 269 270 271  | Next Page >