Search Results

Search found 1083 results on 44 pages for 'revision'.

Page 28/44 | < Previous Page | 24 25 26 27 28 29 30 31 32 33 34 35  | Next Page >

  • svnkit: Problem getting entry name

    - by GreenKiwi
    I'm trying to create a SVN Eclipese EFS plugin and have problems when getting the names of entries. When I make a call to: SVNRepository `//Fetches the contents of a directory into the provided collection object and returns the directory entry itself. SVNDirEntry getDir(String path, long revision, boolean includeCommitMessages, Collection entries)` It correctly returns the entry for the provided path, however, it doesn't set the "name" value on the "returned" entry. Note, the items returned in the collection are all OK. Does anyone know why this is? And/or if there is a workaround? See: http://svnkit.com/javadoc/org/tmatesoft/svn/core/io/SVNRepository.html http://svnkit.com/javadoc/org/tmatesoft/svn/core/io/SVNRepository.html#getDir(java.lang.String, long, boolean, java.util.Collection)

    Read the article

  • Node.js vs PHP processing speed

    - by Cody Craven
    I've been looking into node.js recently and wanted to see a true comparison of processing speed for PHP vs Node.js. In most of the comparisons I had seen, Node trounced Apache/PHP set ups handily. However all of the tests were small 'hello worlds' that would not accurately reflect any webpage's markup. So I decided to create a basic HTML page with 10,000 hello world paragraph elements. In these tests Node with Cluster was beaten to a pulp by PHP on Nginx utilizing PHP-FPM. So I'm curious if I am misusing Node somehow or if Node is really just this bad at processing power. Note that my results were equivalent outputting "Hello world\n" with text/plain as the HTML, but I only included the HTML as it's closer to the use case I was investigating. My testing box: Core i7-2600 Intel CPU (has 8 threads with 4 cores) 8GB DDR3 RAM Fedora 16 64bit Node.js v0.6.13 Nginx v1.0.13 PHP v5.3.10 (with PHP-FPM) My test scripts: Node.js script var cluster = require('cluster'); var http = require('http'); var numCPUs = require('os').cpus().length; if (cluster.isMaster) { // Fork workers. for (var i = 0; i < numCPUs; i++) { cluster.fork(); } cluster.on('death', function (worker) { console.log('worker ' + worker.pid + ' died'); }); } else { // Worker processes have an HTTP server. http.Server(function (req, res) { res.writeHead(200, {'Content-Type': 'text/html'}); res.write('<html>\n<head>\n<title>Speed test</title>\n</head>\n<body>\n'); for (var i = 0; i < 10000; i++) { res.write('<p>Hello world</p>\n'); } res.end('</body>\n</html>'); }).listen(80); } This script is adapted from Node.js' documentation at http://nodejs.org/docs/latest/api/cluster.html PHP script <?php echo "<html>\n<head>\n<title>Speed test</title>\n</head>\n<body>\n"; for ($i = 0; $i < 10000; $i++) { echo "<p>Hello world</p>\n"; } echo "</body>\n</html>"; My results Node.js $ ab -n 500 -c 20 http://speedtest.dev/ This is ApacheBench, Version 2.3 <$Revision: 655654 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking speedtest.dev (be patient) Completed 100 requests Completed 200 requests Completed 300 requests Completed 400 requests Completed 500 requests Finished 500 requests Server Software: Server Hostname: speedtest.dev Server Port: 80 Document Path: / Document Length: 190070 bytes Concurrency Level: 20 Time taken for tests: 14.603 seconds Complete requests: 500 Failed requests: 0 Write errors: 0 Total transferred: 95066500 bytes HTML transferred: 95035000 bytes Requests per second: 34.24 [#/sec] (mean) Time per request: 584.123 [ms] (mean) Time per request: 29.206 [ms] (mean, across all concurrent requests) Transfer rate: 6357.45 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.2 0 2 Processing: 94 547 405.4 424 2516 Waiting: 0 331 399.3 216 2284 Total: 95 547 405.4 424 2516 Percentage of the requests served within a certain time (ms) 50% 424 66% 607 75% 733 80% 813 90% 1084 95% 1325 98% 1843 99% 2062 100% 2516 (longest request) PHP/Nginx $ ab -n 500 -c 20 http://speedtest.dev/test.php This is ApacheBench, Version 2.3 <$Revision: 655654 $> Copyright 1996 Adam Twiss, Zeus Technology Ltd, http://www.zeustech.net/ Licensed to The Apache Software Foundation, http://www.apache.org/ Benchmarking speedtest.dev (be patient) Completed 100 requests Completed 200 requests Completed 300 requests Completed 400 requests Completed 500 requests Finished 500 requests Server Software: nginx/1.0.13 Server Hostname: speedtest.dev Server Port: 80 Document Path: /test.php Document Length: 190070 bytes Concurrency Level: 20 Time taken for tests: 0.130 seconds Complete requests: 500 Failed requests: 0 Write errors: 0 Total transferred: 95109000 bytes HTML transferred: 95035000 bytes Requests per second: 3849.11 [#/sec] (mean) Time per request: 5.196 [ms] (mean) Time per request: 0.260 [ms] (mean, across all concurrent requests) Transfer rate: 715010.65 [Kbytes/sec] received Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 0.2 0 1 Processing: 3 5 0.7 5 7 Waiting: 1 4 0.7 4 7 Total: 3 5 0.7 5 7 Percentage of the requests served within a certain time (ms) 50% 5 66% 5 75% 5 80% 6 90% 6 95% 6 98% 6 99% 6 100% 7 (longest request) Additional details Again what I'm looking for is to find out if I'm doing something wrong with Node.js or if it is really just that slow compared to PHP on Nginx with FPM. I certainly think Node has a real niche that it could fit well, however with these test results (which I really hope I made a mistake with - as I like the idea of Node) lead me to believe that it is a horrible choice for even a modest processing load when compared to PHP (let alone JVM or various other fast solutions). As a final note, I also tried running an Apache Bench test against node with $ ab -n 20 -c 20 http://speedtest.dev/ and consistently received a total test time of greater than 0.900 seconds.

    Read the article

  • How to make teamcity's svn checkout more verbose?

    - by Benju
    We have a very large svn external containing about 30,000 500k files. This checkout can take a long time and we would like to see the progress in the TeamCity logs as it happens. Is there a way to use a more verbose logging when doing the svn checkout than just.... [19:26:00]: Updating sources: Agent side checkout... [19:26:00]: [Updating sources: Agent side checkout...] Will perform clean checkout. Reason: Checkout directory is empty or doesn't exist [19:26:00]: [Updating sources: Agent side checkout...] Cleaning /opt/TeamCity/buildAgent/work/937995fe3d15f1e7 [19:26:00]: [Updating sources: Agent side checkout...] VCS Root: guru 6 trunk with externals [19:26:00]: [VCS Root: guru 6 trunk with externals] revision: 6521_2010/04/27 19:25:58 -0500

    Read the article

  • jquery stop image slider when accordian menu is closed

    - by Nik
    http://jsbin.com/emoba5/3/edit click here to view demo This accordion menu/image slider works well except - When the page loads and when all accordion menu items are closed I need it to show a single image on right. At the moment the page loads a blank picture area and when an accordion menu item is closed it just restarts that menu items picture slider. (So on page load and when all accordions are closed I need it default to an image on the right.) Some of the items only have 1 picture i.e. posture 3 you will notice when you click on it, it shows the image and then goes blank, can it just fade the image in and leave it there if there is only 1 image? Thanks to aSeptik for his help so far, feel free to make a revision of the posted code. Thanks guys Nik

    Read the article

  • Bamboo Versioning

    - by reddy
    Hello guyz, I have a situation where i need to maintain version information of my builds. By googling i found limited information. one way is to create a version file on source control and keep updating. other is to use the source control revision number. final one is to use bamboo build number. i haven't implemented anyone of this before. colud anyone point out the pros and cons of each method. Thank you, Reddy. Please atleast tell me which method have u used to implement the same. Thnq..

    Read the article

  • Non-existent file in limbo prevents push to remote branch (Bazaar VCS)

    - by das_weezul
    Hi! I use Bazaar VCS to version files locally on my notebook. When im in the office I merge the changes to a repository on a windows share and also push all the files there (for backup reasons). My Problem: The last push resulted in an error, because I added a file with a very long filename (I had that problem before ... python doesn't like long filenames). So I removed the file (I didn't need it anyway) and forgot about the problem for a while, because commiting still worked fine. The next time I wanted to push my new revision I got a new error: bzr: ERROR: [Error 3] Das System kann den angegebenen Pfad nicht finden: u'//path/to/remote/branch/.bzr/checkout/limbo/new-8/loooooooongfilename.xls' translation: bzr: ERROR: [Error 3] The system can't find the following path: What I've tried: Deleting the limbo folder-- limbo folder doesn't exist Create the missing path with a dummy-file -- bazaar locks the branch -- unlock -- same problem as before bzr check -- Everything is fine -- No success bzr reconcile -- No success Thanks for reading ;o)

    Read the article

  • Why is this an invalid Turing machine?

    - by Danny King
    Whilst doing exam revision I am having trouble answering the following question from the book, "An Introduction to the Theory of Computation" by Sipser. Unfortunately there's no solution to this question in the book. Explain why the following is not a legitimate Turing machine. M = { The input is a polynomial p over variables x1, ..., xn Try all possible settings of x1, ..., xn to integer values Evaluate p on all of these settings If any of these settings evaluates to 0, accept; otherwise reject. } This is driving me crazy! I suspect it is because the set of integers is infinite? Does this somehow exceed the alphabet's allowable size? Thanks!

    Read the article

  • Ubuntu | Ruby - Can't Remove Ruby 1.9.2 from system

    - by JWally
    I'm new to ubuntu, and might have had a couple of false starts updating ruby / rails. I'm trying to start clean by removing all versions of every thing ruby from my system I think I've installed everything through either Synaptic, or aptitude purge, but I can still do the following: jwally@jwally-laptop:~$ ruby -v ruby 1.9.2p0 (2010-08-18 revision 29036) [i686-linux] jwally@jwally-laptop:~$ gem -v 1.3.7 jwally@jwally-laptop:~$ irb irb(main):001:0> RUBY_VERSION => "1.9.2" but when I type "dpkg -l | grep ruby, I only get the following: rc ruby1.8-elisp 1.8.7.249-2 ruby-mode for Emacsen thanks in advance

    Read the article

  • Rails3 and `cd somehwere && do something`

    - by Samer Abukhait
    I have a rails project that has other projects under it, sub-projects have rake and bundler files. When I do ruby -e `cd sub-project && rake`, or ruby -e `cd sub-project && bundle`, commands work as expected and use the sub-project rake/bundler files. However, when I do the same thing from a Rails3 console (rails 3.0.3), rake gives the error no such file to load -- initializer, and bundle operates as if it was fired from the root directory. I tried the same commands from a Rails2.3.10 console and they worked as expected. Is Rails3 doing something wrong here? I am using Ruby 1.9.2 via RVM. $ ruby -v ruby 1.9.2p136 (2010-12-25 revision 30365) [i686-linux]

    Read the article

  • How to create trunk directory in existing svn repo without a trunk, and move all files to new trunk?

    - by stupakov
    Hi all, I've got an svn repo with no trunk dir. I'd like to create the trunk and branches dirs, and move all my files (currently in the root) into trunk. Creating trunk works just fine: macbook[601] # svn mkdir trunk A trunk macbook[602] # svn commit trunk Adding trunk Committed revision 67. moving everything into trunk fails: macbook[604] # svn move * trunk svn: Cannot copy path 'trunk' into its own child 'trunk/trunk' is there a way to do this without giving it all the file and dir names individually? There are threads describing how to do this with TortoiseSVN, but i'd like to know the right way to do it with the svn commandline. Thanks.

    Read the article

  • In Subversion, I know when a file was added, what's the quickest way to find out when it was deleted

    - by Eric Johnson
    OK, so suppose I know I added file "foo.txt" to my Subversion repository at revision 500. So I can do a svn log -v http://svnrepo/path/foo.txt@500, and that shows all the files added at the same time. What's the fastest way to find when the file was deleted after it was added? I tried svn log -r500:HEAD -v http://svnrepo/path/foo.txt@500, but that gives me "path not found" - perhaps obviously, because the file "foo.txt" doesn't exist at "HEAD". I can try a binary search algorithm going forward through revisions (and that would certainly be faster than typing this question), but is there a better way?

    Read the article

  • How do i serve shortcuts/.lnk from web servers?

    - by acidzombie24
    without using a database i wanted a file to point to the newest revision of a file. Someone suggested using a shortcut. Knowing i can rewrite file.ext to file.ext.lnk i thought it was a great idea. Then i tried it, my server (VS 2010rc) serves the shortcut rather then the file. Not what i wanted... How do i serve the file the shortcut is pointing to? NOTE: I am planing to use windows 2008 as my server so a solution should work on that as well. The OS i am running is windows 7.

    Read the article

  • svn reintegrate a branch with externals fails

    - by dnndeveloper
    using svn 1.6.6 with tortoisesvn 1.6.6 what I am doing: 1) Apply external properties to a folder in the trunk (both single file and folder external, externals are binary files) 2) Create a branch from the trunk and update the entire project 3) Modify a file on the branch and commit the changes, then update the entire project. 4) Merge - "Reintegrate a branch" when I get to the last screen I click "test merge" and get this error: Error: Cannot reintegrate into mixed-revision working copy; try updating first I update the entire project and still the same error. other observations: If I "Merge a range of revisions" everything works fine. If I remove the externals everything works fine using either "Merge a range of revisions" or "Reintegrate a branch" Anyone else having this issue?

    Read the article

  • Engine finish() causes segmentation fault

    - by Becky
    Hello All - I am using M2Crypto revision 723 from the repository. I am trying to clean up my engine. If I have the pkcs11.finish() line in my script, the script finishes but gets a segmentation fault at the end. Without the finish() line, no segmentation fault occurs. Is there something wrong with the way I'm using finish()? dynamic=Engine.load_dynamic_engine("pkcs11","/usr/local/ssl/lib/engines/engine_pkcs11.so") pkcs11 = Engine.Engine("pkcs11") pkcs11.ctrl_cmd_string("MODULE_PATH", "/usr/lib/libeTPkcs11.so") pkcs11.init() # next few steps which I deleted pass password and grab key & cert off token pkcs11.finish() Engine.cleanup() Thanks!

    Read the article

  • Unable to commit to Subversion

    - by Ewan Makepeace
    I have a client who had to rebuild his automated build server. He checked out his project folder from my subversion server but is now no longer able to commit - he gets this error: Error: Commit failed (details follow): Error: Cannot write to the prototype revision file of transaction '551-1' because a Error: previous representation is currently being written by another process Finished!: I have searched Google but although this error has been often reported there is no clear explanation - does anyone on StackOverflow have a solution? UPDATE: Nobody else commits to that repository, so it was not a transaction stuck (at least not from another user). In the end we found that permissions were not set correctly. Not that you would know it from this message, but that fixed the problem.

    Read the article

  • Problem with truncation of floating point values in DBSlayer.

    - by chrisdew
    When I run a query through DBslayer http://code.nytimes.com/projects/dbslayer the floating point results are truncated to a total of six digits (plus decimal point and negative sign when needed). { ... "lat":52.2228,"lng":-2.19906, ... } When I run the same query in MySQL, the results are as expected. | 52.22280884 | -2.19906425 | Firstly, am I correct in identifying DBSlayer as the cause of this effect? (Or the JSON library it uses, etc.) Secondly, is this floating point precision configurable within DBSlayer? Thanks, Chris. P.S. Ubuntu 9.10, x86_64 Path: . URL: http://dbslayer.googlecode.com/svn/trunk Repository Root: http://dbslayer.googlecode.com/svn Repository UUID: 5df2be84-4748-0410-afd4-f777a056bd0c Revision: 65 Node Kind: directory Schedule: normal Last Changed Author: dgottfrid Last Changed Rev: 65 Last Changed Date: 2008-03-28 22:52:46 +0000 (Fri, 28 Mar 2008)

    Read the article

  • Using GIT with Joomla

    - by kwhohasamullet
    Hi Guys, We are a web design company going down the road of setting up a revision management system and all the processes around how we are going to use it etc. We mainly develop our websites on a Joomla and what i wanted to know is how other companies manage their repositories when dealing with a CMS. We mainly deal with template building as well as occasionally customising components/plugins we installed. My main questions are: -Is the best way to store all files (including the Joomla files) in the repo or just files that you make/ change yourself? -Do you keep a copy of the Database somewhere to account for database changes Thanks in advance

    Read the article

  • Difference between Revert and Update in Mercurial

    - by Edan Maor
    I'm just getting started with Mercurial, and I've come across something which I don't understand. I made changes to several files, and now I want to undo all the changes I made to one of them (i.e. go back to my last commit for one specific file). As far as I can see, the command I want is revert. In the page I linked to, there is the following statement: This operation however does not change the parent revision of the working directory (or revisions in case of an uncommitted merge). To undo an uncomitted merge, you can use "hg update -C -r." which will reset the parents to the first parent. I don't understand the difference between the two (hg revert vs. hg update -C -r). Can anyone enlighten me as to the difference? And in my case, do I really want the revert or the update to go get rid of the changes I made to the file? Thanks

    Read the article

  • svn cat for git

    - by sanxiyn
    I am looking for the equivalent of svn cat in git. Yes, I am aware that the similar question was asked here. The answer is to use git show rev:path. However, svn cat can be used for the remote repository. That is, I can do svn cat url@rev and get the file from the specified revision of the remote repository, without getting the whole repository. My understanding is that git show only applies to the local repository. A workaround I found is to use gitweb interface to get the blob.

    Read the article

  • Redmine connecting to SVN through SSL

    - by Pekka
    I am having trouble connecting Redmine to a locally hosted subversion repository using SSL. I suspect it's the self-signed certificate that usually triggers a warning in the SVN client and browser. When I try to connect to the local repo through SSL in Redmine, I get a red "Revision not available" error. When I try connecting through svn://, the connection times out, and I have to restart the web server. Connecting without SSL works without problems. It would be nice to run subversion on SSL to make it safely accessible from the outside as well. I could run the repository through plain HTTP but would like SSL for outside communication. As far as I understand, subversion can't be run both ways at the same time. Does anybody know what to do in such a situation? Is there a configuration setting to ignore invalid certificates somewhere?

    Read the article

  • Using Visual Studio for open source projects (projects management & legal issues)

    - by Goofy
    Hello! I'm working on small open source project for developed Windows. I created new project using Visual Studio wizard and now I want to publish it using subversion and develop it with friends. What projects' configuration files I have to add to repository to make my friends able to download current project's revision and open it in Visual Studio? List item University where I study is a member of Microsoft Academy Alliance project, which make students able to download legal copies of many Microsoft's products. Can I use this copy to develop open source project and publish created code in Internet?

    Read the article

  • How to keep history of record updates in MySQL?

    - by Proxium
    I have to create a code in PHP that will allow me to keep the history of record updates in MySQL database so I can find by date an old revision. Here is the example of what I actualy want to achive: http://en.wikipedia.org/w/index.php?title=Tunisia&action=history The data are mostly Numbers that we record about the company for generating reports and to extract indices. I plan to use codeigniter for it's simplicity and I'm looking for idea about a framework or an opensource project that use the same approche to keep history of modifications in the database.

    Read the article

  • What software development process do you use and how do you implement it?

    - by clyfe
    Post only what you do use not what you would like to use, so we can see what is the most popular in real life. I am interested only in theese issues: Project Model (waterfall, agile...) How are requirements gathered (and stored)? Revision control - what software, what workflow Build automation, what software, where does it fit ? How is the testing done ? How is the documentation done ? How is the quality assurance done ? Please provide short objective answers, don't speak from the books. EXAMPLE: In my company we are a small team of 5 people and we develop webapps using ruby. agile PM cucumber requirements git SCM - Integration Manager Workflow integrity CI rspec automated tests the project lead creats the documentation skeleton then it is filled by the developers ensure quality by peer reviewing code and manual peer-testing

    Read the article

  • Some basic COM question...

    - by smwikipedia
    I have just finished my first COM server DLL. And it runs smoothly. So I'd like to show my understanding for now and hear your critics. 1- How COM simply works? COM - "The Call Chain" COM Lib methods - Traditional DLL exports - Classes encapsulated in the COM DLL 2- With C++, the benefits like "interface" in OOP can only be taken advantage of at the source level. With COM, these benefits can be used at a binary level. 3- Some illustration about interface &pInterface ------- pInterface ---------- Interface----------------- methods Ixx ** Ixx * (method table) (void **) A Interface is a data structure in memory. It's nothing but a memory area containg a method table. Is my understanding alright? Thanks for your revision.

    Read the article

  • Is it really wrong to version documents using CouchDB's default behaviour?

    - by Tomas Sedovic
    This is one of those "I know I shouldn't do this but it's oh so convenient." questions. Sorry about that. I plan to use CouchDB for storing a bunch of documents and keeping their entire revision history. CouchDB does the versioning automatically, but it is strongly discouraged for programmer's use: "You cannot rely on document revisions for any other purpose than concurrency control." From what I've found on the CouchDB wiki, the versions can get deleted either during compaction or during replication. As far as I can tell, Compaction must always be triggered manually and Replication occurs only when there's more than one database server. The question is: if I won't run compaction and will use only single database instance for my documents, can I just use CouchDB's document versioning and expect it to work? What other problems I might run into? E.g. does not running compaction hurt the performance or consume significantly more disk space (than if I did handle the versioning manually)?

    Read the article

< Previous Page | 24 25 26 27 28 29 30 31 32 33 34 35  | Next Page >