Search Results

Search found 4296 results on 172 pages for 'git clone'.

Page 65/172 | < Previous Page | 61 62 63 64 65 66 67 68 69 70 71 72  | Next Page >

  • Duplicating an array of strings.

    - by Jon
    arr = ["red","green","yellow"] arr2 = arr.clone arr2[0].replace("blue") puts arr.inspect puts arr2.inspect produces: ["blue", "green", "yellow"] ["blue", "green", "yellow"] Is there anyway to do a deep copy of an array of strings, other than using Marshal as i understand that is a hack. I could do: arr2 = [] arr.each do |e| arr2 << e.clone end but it doesn't seem very elegant, or efficient. Thanks

    Read the article

  • How do I get git whatchanged to show a combined list of files that have changed?

    - by Chirag Patel
    I ran the following comand git whatchanged 7c8358e.. --oneline and got the below output. Is there a way to generate a single combined list of files that changed across all commits? In other words, I don't want files to show up more than once in the below list. Thanks! 4545ed7 refs #2911. error on 'caregivers_sorted_by_position' resolved in this update. it came up randomly in cucumber :100644 100644 d750be7... 11a0bd0... M app/controllers/reporting_controller.rb :100644 100644 7334d4d... e43d9e6... M app/models/user.rb e9b2748 refs #2911. group dropdown filters the list to only the users that belong to the selected group :100644 100644 fc81b9a... d750be7... M app/controllers/reporting_controller.rb :100644 100644 aaf2398... f19038e... M app/models/group.rb :100644 100644 3cc3635... 7a6b2b1... M app/views/reporting/users.html.erb 48149c9 refs #2888 cherry pick 2888 from master into prod-temp :100644 100644 3663ecc... f672b62... M app/controllers/user_admin_controller.rb :100644 100644 aaf2398... 056ea36... M app/models/group.rb :100644 100644 32363ef... bc9a1f2... M app/models/role.rb :100644 100644 91283fa... 7334d4d... M app/models/user.rb :100644 100644 d6393a0... bae1bd6... M app/views/user_admin/roles.html.erb 994550d refs #2890. all requirements included. cucumber has 1 exception in bundle_job for count of data rows. everything else green :100644 100644 145122d... 869a005... M app/controllers/profiles_controller.rb :100644 100644 f1bfa77... 2ed0850... M app/views/alerts/message.html.erb :100644 100644 e9f8a34... f358a74... M app/views/call_list/_item.html.erb :100644 000000 fda1297... 0000000... D app/views/call_list/_load_caregivers.erb :000000 100644 0000000... fda1297... A app/views/call_list/_load_caregivers.html.erb :100644 100644 168de9e... 43594f4... M app/views/call_list/show.html.erb :100644 100644 e178d7f... 0fe77e1... M app/views/profiles/edit_caregiver_profile.html.erb 7396ff6 refs #2890. fixed --we're sorry-- error :100644 100644 d55d46d... fc81b9a... M app/controllers/reporting_controller.rb 7c8358e refs #2897 link on online store back to http://www.halomonitoring.com :100644 100644 d6f94f4... 8bc9c52... M app/views/orders/new.html.erb

    Read the article

  • How to be Agile when new work keeps affecting completed work?

    - by jdln
    The project I'm working on is to re-skin an existing website. The functionally will stay the same, its just the styles that are changing. The HTML is not changing, I'm only modifying the CSS files. The site is pretty complex. There are dozens of pages. Users can be logged in and have a number of different roles. Depending on their role the content of the page and what pages they are allowed to see varys. We're using GIT and Github. I'm trying to write CSS that works as components. So when the same form elements, headings, etc appear on multiple pages they are already styled and are consistent. Most of time this is working well. Sadly the format and class names in the HTML are at times messy and unpredictable. When I fix something on one page it can break another. The job is also harder as no one knows exactly all the variations that are possible due to the user roles. As such I'm continuously finding new variations as I go along. I'm making headway by putting a lot of comments in my CSS. If I need to remove a CSS rule Ill comment it out so I can still see it with the chrome dev tools, and ill put a comment in the CSS saying why I removed it and for what page this was done. This means that if on another page I'm about to add add the rule to fix a different problem, there is more of a chance I will see how this would break the first page. This allows me to either find a different solution that will work for both pages, or I can make the override page specific. This has been working quite well for me. If I had complete free reign and the only deadline was to finish the project by the end then this method would be fine. However my manager is trying to mitigate risk by breaking the work into areas to be completed per sprint. This is counter to how I have been approaching things as something like my typography styles will affect all other pages on the site. The other issue is that the different stakeholders want to sign off each section as I go along. However once I've finished a section it may change if I change CSS that affects it and also affects a new section I'm working on. I've asked that the stakeholders have a quick unofficial sign off in stages (eg per sprint), and have the final official sign off at the end of the project, but this is being met with resistance. I do understand why it would be higher risk to do this, but the only way to guarantee that a signed off section will not change is to make ALL future changes page specific. In addition to this I'm being told that all work that I push to the Git repo should be ready to go live, and as such should not contain any code comments. This is risky for me as I wont know until I've finished the site if I will ever benefit from these comments or not. Has anyone else been in a similar situation and managed to find a compromise that worked for my development approach and also the desires of management and stakeholders to have a more Agile approach? A more Agile workflow works great when you can break the work into components and know that once something is done it wont be affected by future work. However the nature of this project makes this hard to achieve.

    Read the article

  • Vertex 2 SSD is running faster than my Vertex 3 SSD?

    - by Kairan
    I used Acronis Disk Director to do a direct clone of my C:\ windows 7 x64 drive from my Vertex 2 to my new Vertex 3 SSD (Just to show the drive software winstall everything is identical.) I ran a performance test on Windows using the Windows Experience Index. The rating I am receiving when booting on the Vertex 2 is 7.5 While I am getting only a rating for the Vertex 3 of 6.9 My understanding is that the read/write speeds of the Vertex 2 is only up to 250MB/sec while the Vertex 3 is up to 500MB/sec. Copying a single file (3GB in size) from the Vertex 3 to itself was getting speed of approx 70-80MB/sec This speed is no better (maybe worse) than what I got from the Vertex 2 I am connected via the SATA 3 port on the motherboard, using an SATA 3 cable Is this issue caused by the drive cloning? Do I have a bad SSD?

    Read the article

  • How do I make a VMDK format virtual hard drive from a physical partition on a disk?

    - by Ahmad
    I have a 320 GB HDD, which actually only has an 80 GB NTFS format partition which was being used by a Windows 7 system ... I want to create a VMDK format clone of this partition, so that I can use it with VMware .. However, tradition VMDK creation programs normally make a VMDK for an entire disk, whereas I just want to make a VMDK for the one 80 GB partition ... This is important because the other 240 GB on the physical source HDD is just unallocated area, and including that in a VMDK file is just a plain waste of space .. So how to make a VMDK file for a specific partition ? Any tool available for this ?

    Read the article

  • Will disk cloning resolve bad stripes on RAID?

    - by user13323
    Hi. We have a logical RAID1 drive in bad stripes state, which kept that status even after replacement and rebuilding of both drives, and gives errors in Windows logs about failure of writing to disk. IBM support suggests erasing and re-creating the RAID, then re-installing the Windows. The resulting down-time unacceptible for us, so we want to clone the RAID (via Acronis True Image), erase and re-create the RAID, then dump the cloned data back. Following IBM logic where RAID erasing and re-creation resets the whole RAID meta-data, this should clear the bad-stripes status, and start from a blank page. Question is if such strategy is possible, and will produce the desired effect? Any idea is appreciated - thanks in advance!

    Read the article

  • What does advantages of WIM image that file based vs. ISO or another sector based images in Administrator

    - by saber tabatabaee yazdi
    answers of this question is very important to administrators to choose between Microsoft solution or other 3rd Party Tools. WIM solution is file based and others are sector by sector images like Acronis or other backup and clone tools like Norton Ghost or Cloning in vmware ESXi and etc. My question about the advantages of WIM to edit or deploy in Enterprise? to get agreements of managers to choose between these alternative decisions. Are we had any Key Performance Indicators in Change management? Are there any statistics about compares of these solutions? What Time is spend in any solotion to get image or change images or revert and apply images? Is it possible to change and modify images that made in All of these solutions as easy as in WIM file?

    Read the article

  • Cannot acess the new cloned server even after new IP address assignment

    - by tough
    I was able to clone a Ubuntu 10.04 server residing in Cloud. It appeared that I was not getting some IP for the new VM so I followed some of these: # cd /etc/udev/rules.d # cp 70-persistent-net.rules /root/ # rm 70-persistent-net.rules # reboot I didn't follow the later commands as I was unable to see two eth MACs as available in the referenced site. After this I am able to see some the IP for it, and is different form the original IP, I have added new IP to DNS server. Now when I try to access it with its assigned(new) domain it is directed to the old server. I can see both the VMs running with different IP. Where I might have gone wrong, I am new to this admin thing.

    Read the article

  • Cloning a linux system from sdx to cciss

    - by churnd
    I have an HP ML 310 server running CentOS Linux 5.5. I'm buying a RAID card (LSI 9260-8i) to set up a mirrored OS drive. Right now, the boot drive is set up with GRUB installed on the MBR of /dev/sda & has a 100MB /boot partition for /dev/sda1, then the rest is configured in LVM with a 20GB with a 20GB VG for the root partition & ~80GB VG for home. The new disk sizes will also be slightly larger as well. What is the best way to clone the boot drive to the new CCISS device?

    Read the article

  • Why doesn't Ghost 2003 offer to fill the destination drive?

    - by Neil
    Because it is dangerously low on disk space, I want to upgrade an SBS 2003 server by replacing its existing 72GB drive with a 364GB drive. When I tried to use Norton Ghost 2003 to clone the disk it didn't suggest that I use the entire new drive. I'm worried that I caused the process to fail by overriding its decision - although the cloned drive boots in Safe Mode, if I try booting it normally then none of the SQL Express instances start and something causes the server to reboot before even the Ctrl+Alt+Del screen appears. Does Ghost 2003 know something that I don't? Or should I be using some other software?

    Read the article

  • What's the easiest way to upgrade a single laptop hard drive?

    - by rutherford
    My HD is coming near the end of it's life and I wondered if producing an image of the original was all that was required? What's the best software to use for this and how can I accomplish it - can I use my other laptop for the image file or do I need to buy an external HD to accomplish the clone? Does having several partitions of both FAT32 and NTFS with partition magic installed affect things at all? Is it just a matter of making the image, swapping in the new HD and copying the image to that?

    Read the article

  • Backing up to smaller drive

    - by Dave
    In a few hours I'll have a new 500GB Sony laptop, filled with the usual Sony rubbish which I'll promptly be replacing with Ubuntu or Crunchbang or something. However, first I want to make a full clone of the drive (including recovery partitions), should I wish to return it to Sony or sell it on in its factory state. The problem is that the only backup drives I have are less than 500GB - the biggest I have is 250GB or so! So I need to backup and compress on-the-fly. What's the best way to do this? Presumably dd piped into gzip would do the trick, or does anyone have any other suggestions to accomplish this?

    Read the article

  • Cloning Fresh Windows 7 with external HDD, DVD or CDs?

    - by hhh
    I want to create Windows installation disk (not necessarily CD) from my Windows 7 Pro laptop, not sure what it actually means here. The material is about 50-60GB, requiring about 12 pcs of 5GB disk (not going to happen, too much work to use them later). I have the Windows serial on the laptop bottom and this is so-called firm -laptop (no idea what it actually mean, better warranty and some instant support thing apparently). Now how to do the clone with external material such as DVDs? how can I create mock-windows-installation medium or some real windows-installation medium? I am now not sure whether Windows offers W7 -installations medium online, well my W -peer mentioned some MSDAA -something. Perhaps related Cloning Fresh Windows 7 -fsed HDD to Linux Server because having no external HDD or disks for the Backup

    Read the article

  • Debian/Linux backup files changed by user

    - by verhogen
    I would like to backup my server that is hosting a few websites in such a way that I can restore everything to the way it was from a fresh format. I know that I should backup all the home folders and then probably my /etc/ folders. Is there a way to figure out all the folders that are relevant for backup in that they were not automatically generated or installed from apt-get? It would ideally restore all the users with their current passwords as well. Basically, enough to clone the system but only copying configuration files.

    Read the article

  • OutOfMemoryException with Image.Clone - Only on Windows 2003

    - by Jeff Rapp
    So here's my issue. I have an image that I need to shrink. The original image is a grayscale PNG, which isn't a huge issues except that when I shrink it down, the thermal label printers pickup the artifacts and print them on the label. So, what I did was change the image to black & white (Format1bppIndexed) before resizing, like this: Dim bte() As Byte = System.Convert.FromBase64String(imgStr) Dim ms As New IO.MemoryStream(bte) Dim bmp As New System.Drawing.Bitmap(ms) Dim monoImage As Drawing.Bitmap = New Drawing.Bitmap(1200, 1800, Drawing.Imaging.PixelFormat.Format1bppIndexed) Dim rect As New Drawing.Rectangle(0, 0, 1200, 1800) monoImage = bmp.Clone(rect, Drawing.Imaging.PixelFormat.Format1bppIndexed) And then I resize it. This code works fine on my Windows 7 machine, but when I run it on the Windows 2003 Server box that it calls home, it always throws an OutOfMemoryException when it hits the bmp.Clone line. Any ideas as to what's happening, or perhaps a better solution to converting the image to B&W?

    Read the article

  • Makefile error: Unexpected end of line seen

    - by Winston C. Yang
    Trying to install Git, I ran configure and make, but got the following error message: make: Fatal error in reader: Makefile, line 221: Unexpected end of line seen The Makefile looks like: 218: GIT-VERSION-FILE: FORCE 219: @$(SHELL_PATH) ./GIT-VERSION-GEN 220: -include GIT-VERSION-FILE 221: 222: uname_S := $(shell sh -c 'uname -s 2>/dev/null øø echo not') What's causing the error? The following information may or may not be relevant: I tried to install Git 1.7.0.3 on SunOS 5.9 (Solaris 9) in a directory in my account. The gcc version is 3.4.2 (older then the version of 3.4.6 stated by sunfreeware.com). I don't have root privileges.

    Read the article

  • A file was added to git on commit n. How do I add it instead to commit n-m?

    - by carleeto
    I have a branch. Half way through I noticed git was not tracking a file that it should have been and so I added it as part of a commit and continued with my work. Now, I'm doing a git bisect and all commits before the file was added do not build. So I'm thinking, I need to split the commit that added the file into two parts: the file add and the rest of the commit. I then need to re-order the commits so that the file add commit will be at the beginning of my branch. Is this the correct solution or is there a better way of doing it?

    Read the article

  • Faster way to clone.

    - by AngryHacker
    I am trying to optimize a piece of code that clones an object: #region ICloneable public object Clone() { MemoryStream buffer = new MemoryStream(); BinaryFormatter formatter = new BinaryFormatter(); formatter.Serialize(buffer, this); // takes 3.2 seconds buffer.Position = 0; return formatter.Deserialize(buffer); // takes 2.1 seconds } #endregion Pretty standard stuff. The problem is that the object is pretty beefy and it takes 5.4 seconds (according ANTS Profiler - I am sure there is the profiler overhead, but still). Is there a better and faster way to clone?

    Read the article

  • Is .gitignore not working or I have misunderstood it?

    - by Shubham
    I am very new to git. I have a .gitignore in the my working folder. *.jpg *.gif *.png system/* */Zend/* .idea/*.* Well, I did git init and then git add *. At this it worked fine and ignored the above files. But when I did some changes, ran the same command it puts the ignored files into staging area. The reason why I am using git add * is because I work on many files and adding each file would be a overkill. Update: Here are messages when I run git add * second time.. #new file: application/vendors/Zend/XmlRpc/Value/String.php #new file: application/vendors/Zend/XmlRpc/Value/Struct.php ... The list is too long.

    Read the article

  • How to determine files that are subjected to filter via gitattributes when filter is executed?

    - by rasjani
    I have bunch of ruby scripts in a git repository and it seems to be really hard to enforce people to write properly indented code. I also have a small ruby script that formats to code to specific standard and now i would like to run that as a a filter script so that junk wont get committed into repository. echo "*.rb filter=rubyfilter" > .gitattributes echo "[filter \"rubyfilter\"]" >> .git/config echo " clean = /home/rasjani/bin/rbeauty" >> .git/config echo " smudge = /home/rasjani/bin/rbeauty" >> .git/config does the dirty trick git side but the ruby script should then process the files affected: how / where do i look those up from ?

    Read the article

  • GIT vs. Perforce- Two VCS will enter... one will leave.

    - by Justin Bozonier
    So I'm in the process of getting GIT sold at work. First thing I need is to convince everyone that GIT is better at what they're already used to doing. We currently use Perforce. Anybody else go through a similar sale? Any good links/advice? One of the big wins is that we can worth with it disconnected from the network. Another win imo is the way adds/checkouts are handled. More points are welcome! Also we have about 10-20 devs total.

    Read the article

  • gitk without X11 [closed]

    - by svnpenn
    It has been noted here that Tcl/Tk, and in turn gitk now require X11 under Cygwin. Having run it before and after this change it seems like extreme overkill. I use gitk very lightly, mostly sticking to simply command line git. How could I go about using gitk without X11, perhaps manually installing old version of Tcl/Tk? After some tinkering, I came up with this script that allows gitk without X11 #!/bin/sh # Requires Cygwin packages: git, make, mingw64-i686-gcc-core, wget # Install Tcl wget prdownloads.sf.net/tcl/tcl8.5.12-src.tar.gz tar xf tcl8.5.12-src.tar.gz cd tcl8.5.12/win ./configure --host i686-w64-mingw32 make install cd - # Install Tk wget prdownloads.sf.net/tcl/tk8.5.12-src.tar.gz tar xf tk8.5.12-src.tar.gz cd tk8.5.12/win ./configure --host i686-w64-mingw32 make install cd - # Install gitk cd /usr/local/bin wget raw.github.com/git/git/master/gitk-git/gitk chmod 700 gitk echo 'cygpath -m "$1" | xargs -I% wish85 % -- ${@:3}' > wish cd -

    Read the article

  • Flaws in my PHP development setup - sharing sources causing lags

    - by Wiktor
    I have following development setup for my PHP projects: Working station running on Windows 7 with PhpStorm IDE. GIT for version controlling. CentOS on virtual machine (VirtualBox) with Apache and MySQL (copy of production server). So far, I've been sharing project's source folders between host and guest systems and it was working quite well only really slow. The reason behind this is that Apache was reading files from remote folder (mounted locally). After doing some research, I found out that this set up can be improved by using disk mapping (Samba) instead of folder sharing. So I did that change. I configured my PhpStorm to automatically deploy files to mapped drive. Everything works like a charm now, except for one problem - when I change branches I need to synchronize project's local folder with the one on mapped drive and that takes time, a lot of time (like branching in SVN). Is there another way to handle this than just working on files directly on mapped drive?

    Read the article

  • Keeping up with upstream changes while adding small fixes or even major changes

    - by neo
    Often I need to apply some small fixes (to make them run on my environment) or even change some parts of the software (to tailor it to my needs) to software from outside. However this obviously creates problem with updating said software, even when it changes nothing related to my fix. It would be easier when the software provided integration for some kind of plugins but more often than not it doesn't. What would be an ideal workflow regarding that? Most of the projects are git repos I pulled from outside. How should I apply my changes so that I can update painlessly? You can assume that external changes are much more often and larger than my own ones, so reviewing each one of them won't be a solution.

    Read the article

< Previous Page | 61 62 63 64 65 66 67 68 69 70 71 72  | Next Page >