How can i download all directory from server recursively. I use WinINet and this problem solve very hard. I interested there is some easy way that download files and folders from ftp on visual-c++?
Hi.
I want to use Git to backup my home drive, but I also want to use it as a version control system for projects that will be stored in my home drive.
How would I go about doing that? Do I .gitignore all the projects root folders and make new repositories for them?
I have a memory stream that contains a zip file in byte[] format .
Is there any way I can unzip this memory stream, without any need of writing the file to disk ?
In general I am using ICSharpCode.SharpZipLib.Zip.FastZip to unzip a file ,
But any way to unzip a memory stream ? and store the files in another memorystream or in byte[] format according to the files/folders present in the zip ?
Any way I can use the Memorymapped files feature in this scenario ?
I am trying to download a list of files from a ftp server using:
NSURL *ftpUrl = [NSURL URLWithString:@"ftp://xxx:[email protected]"];
I HAVE to use FTP, so please do not recommend an alternative. I want to know how to get the list of files from the root directory and also navigate inside folders (of FTP server).
P.S: I do not think external library would be required to accomplish such an easy task, but if I must, any recommendations would be appreciated.
Thank You :)
i use a treeview to display files and folders like Windows Explorer. it has a NodeMouseClick event but sometimes when i click +, this event doesn't fire.
private void treeView1_NodeMouseClick(object sender, TreeNodeMouseClickEventArgs e)
{
MessageBox.Show("node mouse click");
}
can anyone explain for me why ? and how to know whenever i click + ?
thanks in advance!
(Assuming the network folders/permissions are correctly set up and working in Windows, and a 'default' PHP setup...)
Is it possible to use UNC network paths [like \\ServerName\Folder\file.txt] in PHP's functions like file_get_contents(), fopen(), etc?
And/or, what special cases allow/disallow this?
Hi,
Need some help gathering thoughts on this issue.
Our team is moving ahead with the idea that separating the authenticated and public sections of our app in two separate folders will allow us to be more organized and secured.
I have seen this approach for Admin apps within the site but never for authentication.
We are currently using Authlogic.
What would be the disadvantage of this?
Thanks for your help.
Is there a way to create a working copy with all child folders but with out files in it.
My repo is very big, I only need the empty directory structure to be created.
I'm making a game using C++ and DirectX and it'd be nice to share my groups progress with other people by providing a single executable file instead of having to run a setup wizard.
I've checked my Release and Debug folders but those executable don't work. The release .exe is something like 21KB which doesn't seem right.
So I guess what I'm asking is how to make a single executable with all of the libraries and resource files included.
I get this message at runtime of ASP.NET 2 page :
The page 'MyFolder/blabla.aspx' cannot use the user control 'MyFolder/MyControl.ascx', because it is registered in web.config and lives in the same directory as the page.
Of course I can separate them to 2 different folders and thus solve the problem, but the question is :
WTF !?!?! Why I can't put them in the same folder ?!
Why can't they all .. get along !?! :)
Thanks
I need that works at least with svn, that reads your svn repository and displays information about files and folders contained in it as well as maybe revision info too.
I see so many solution , but none of them was matching my problem .
As i am new to iOS development .
I have a String
abc/_api/web/GetFolderByServerRelativeUrl('Documents/Root Folder/Level 1/Level 2/Level 1-1/Level 1/')/folders.
I want to remove Remove Substring after last "/"
for example my new string should look like this.
abc/_api/web/GetFolderByServerRelativeUrl('Documents/Root Folder/Level 1/Level 2/Level 1-1/Level 1/)
Can someone suggest me the best way to achieve this , thanks ,
Hi,
I'm looking for a way to zip an entire directory, with an unknown number of files and folders. Is there any way I can do this by passing the name of the top-level folder?
Thank you.
Hi, i'm not an expert of Nant, so i'll have to ask this redicolus question.
i have a variable called svn.source.root which point to c:\folderA\FolderB\FolderC
how can i make svn.source.root.modified variable to point to 2 folders up? i.e, folderA
Obviously, the following didn't work:
please help.
thanks
A recent issue has arised.
I have my local copy of code at ~/code/ directory where I made a git init. Whenever I copied a folder with files within it, git recognized folders and files normally.
But today I copied a folder and when I did a git status, git recognized hw2/ folder as a file.
Additionally, I removed .git folder and reinitialized git and now every folder is recognized as file.
Any ideas why this could be?
Every now and then I have a problem with SVN inside eclipse folder gets locked, I have to check out projects few times update and stuff like that . is there a SVN that I can use to commit files directly from console or windows folders?
hi,
I'm using path auto module to automatically stores images in the projects subfolder. I'm using the [title] token and I had a unpleasent surprise: the title "abc / dce", creates 2 folders.
I was not expecting this because in the pattern descrition, [title] is clearly distinguished by [title-raw].
[title] Node title
[title-raw] Unfiltered node title. WARNING - raw user input.
So, how can I fix this issue, I would like to use the title I have in the url "abc-dce"
thanks
StackOverflow users suggested I try my question here.
I'm using Server 2008 EE and IIS 7. I've got a site that I've migrated over from XP Pro using IIS 5.
On the old system, I was using IIS Password to use simple .htaccess files to control a couple of folders that I didn't want to be publicly viewable.
Now that I'm running a full-blown DC with a more powerful version of IIS, I decided it'd be a good idea to start using something slightly more sophisticated. After doing my research and trying to keep things as cheap as possible with a touch of extra security, I decided that Digest Authentication would be the best way to go.
My issue is this:
With Anon access disabled and Digest enabled, I am never prompted for credentials.
when on the server, viewing domain[dot]com/example will simply show my 401.htm page without prompting me for credentials.
when on a different network/computer, viewing domain[dot]com/example again shows my 401.htm without prompting for credentials.
At the site level I only have Anon enabled. Every subfolder, unless I want it protected, has just Anon enabled. Only the folders I want protected have Anon disabled and Digest enabled.
I have tried editing the bindings to see if that would spark any kind of change... www.domain.com, domain.com, and localhost have all been tried. There was never a change in behavior at any permutation (aside from the page not being found when I un-bound localhost to the site).
I might have screwed up when I deleted the default site from IIS. I didn't think I'd actually need it for anything, but some of what I have read online is telling me otherwise now.
As for Digest settings, I have it pointed to local.domain.com, which is the name assigned to my AD Domain. I'm guessing that's right, but honestly have no clue about what a realm actually is.
Would it matter that I have an A record for local.domain.com pointing to my IP address?
I had problems initially with an absolute link for 401.htm pages, but have since resolved that. Instead of D:\HTTP\401.htm I've used /401.htm and all is well. I used to get error 500's because it couldn't find the custom 401.htm file, but now it loads just fine.
As for some data, I was getting entries like this from access logs:
2009-07-10 17:34:12 10.0.0.10 GET /example/ - 80 - [workip] Mozilla/4.0+(compatible;+MSIE+7.0;+Windows+NT+5.1;+.NET+CLR+1.1.4322;+.NET+CLR+2.0.50727;+InfoPath.2) 401 2 5 132
But after correcting my 401.htm links now get logs like this:
2009-07-10 18:56:25 10.0.0.10 GET /example - 80 - [workip] Mozilla/5.0+(Windows;+U;+Windows+NT+5.1;+en-US;+rv:1.9.0.11)+Gecko/2009060215+Firefox/3.0.11 200 0 0 146
I don't know if that means anything or not. I still don't get any credential challenges, regardless of where I try to sign in from ( my workstation, my server, my cellphone even ). The only thing that's seemed to work is viewing localhost and I donno what could be preventing authentication from finding it's way out of the server.
Thanks for any help!
Jon
According to Intel and Dell, by board is technically a 'desktop' board and they therefore do not support Intel USB 3.0 eXtensible Host Controller drivers for Windows Server 2008 (R2 x64). I'm trying to find a workaround. I found an entry on someone who tried to tackle this, but I can't make his fix work for me. Below, I have copied both his entry, and my reply. I'm a loyal stackoverflow user, and hopefully the people here at serverfault can help me:
anyforumuser Re: GA-Z77X-UD5H USB3 Drivers not installing?
« Reply #6 on: July 05, 2012, 04:12:59 am »
Thanks to JoeMiner , his process for the network drivers gave me the clues to figure out to get the USB3 drivers working.
I have got the intel USB3 drivers working at full speed in win server 2008r2
you have to edit the following file :
1. mup.xml
in change the "Windows7" to "W2K8"
2. in setup.if2
under [groups]
line starting with "HSCSDRIVER "
change the "IsOS( ... )" entry to "IsOS(WIN2008_R2,WIN2008_R2_MAXSP)"
inf files
for all copy the content of the [Intel.NTAMD64.6.1] group to the [Intel.NTAMD64.6.2] group
driver folders.
here i am not entirely sure which is correct so there are some double up's.
in the drivers folder copy the "Win7" folder to "win2008" , "win2008_r2" and "x64"
ie your drivers folder should now contain the "win2008" , "win2008_r2" and "x64" folders and they contain contents of the win7 folder (the inf files should of already been fixed)
Run install , It should install properly and work now.
You will have to reboot
If it doesn't work remove the intel usb3 controllers from device manager and get it to "scan for hardware changes"
Good luck !!!
benevida Re: GA-Z77X-UD5H Intel Network Drivers not installing?
« Reply #7 on: August 13, 2012, 02:21:14 pm »
Thank you anyforumuser! A process for getting this driver installed was exactly what I needed.
However, I've hit a snag. I believe I've followed every step exactly as written, but I'm getting an error during installation. I get the message "One or more files that are required for installation are either missing or corrupted. Setup will exit." Behind the error, the 'Setup Progress' shows the current step as "Copying File: C:\Program Files (x86)\Intel\Intel(R) USB 3.0 eXtensible Host Controller Driver\Drivers\iusb3xhc.man". I've checked the installation files, and iusb3xhc.man seems to be a viable file in all of the Windows 2008 sub-directories of the Drivers folder. Therefore I don't see how the file could be missing and I doubt that it is corrupted, (although it does NOT exist in the \Drivers\HCSwitch folder). I opened 'Setup.if2', and there are two aspects to the step of copying iusb3xhc.man that caught my eye. First, the steps immediately preceding are set to 'error=ignore'. If they hadn't completed successfully, this is the first step where we'd hear about it. Second, this is the first step where the relative path '%source%\drivers\%_os%\%_ia%\' is used. If I haven't named the Windows 2008 sub-directories correctly, I could see where things are fouling up.
In any event, if someone could take a look and make suggestions I'd appreciate it. Thank you.
On Mac OS X 10.9.3
Picked a box from the VagrantCloud
Init the vagrant box
$ vagrant init chef/freebsd-9.2-i386
A `Vagrantfile` has been placed in this directory. You are now
ready to `vagrant up` your first virtual environment! Please read
the comments in the Vagrantfile as well as documentation on
`vagrantup.com` for more information on using Vagrant.
List the files
$ ls -al
-rw-r--r-- 1 joel staff 4831 Jun 5 17:17 Vagrantfile
Vagrantfile content
VAGRANTFILE_API_VERSION = "2"
Vagrant.configure(VAGRANTFILE_API_VERSION) do |config|
config.vm.box = "chef/freebsd-9.2-i386"
end
Starting my virtual box leads to Errors
$ vagrant up
Bringing machine 'default' up with 'virtualbox' provider...
==> default: Box 'chef/freebsd-9.2-i386' could not be found. Attempting to find and install...
default: Box Provider: virtualbox
default: Box Version: >= 0
==> default: Loading metadata for box 'chef/freebsd-9.2-i386'
default: URL: https://vagrantcloud.com/chef/freebsd-9.2-i386
==> default: Adding box 'chef/freebsd-9.2-i386' (v1.0.0) for provider: virtualbox
default: Downloading: https://vagrantcloud.com/chef/freebsd-9.2-i386/version/1/provider/virtualbox.box
==> default: Successfully added box 'chef/freebsd-9.2-i386' (v1.0.0) for 'virtualbox'!
==> default: Importing base box 'chef/freebsd-9.2-i386'...
==> default: Matching MAC address for NAT networking...
==> default: Checking if box 'chef/freebsd-9.2-i386' is up to date...
==> default: Setting the name of the VM: freebsd92-i386_default_1401982167145_49633
==> default: Fixed port collision for 22 => 2222. Now on port 2201.
==> default: Clearing any previously set network interfaces...
==> default: Preparing network interfaces based on configuration...
default: Adapter 1: nat
==> default: Forwarding ports...
default: 22 => 2201 (adapter 1)
==> default: Booting VM...
==> default: Waiting for machine to boot. This may take a few minutes...
default: SSH address: 127.0.0.1:2201
default: SSH username: vagrant
default: SSH auth method: private key
default: Warning: Connection timeout. Retrying...
default: Warning: Connection timeout. Retrying...
==> default: Machine booted and ready!
Sorry, don't know how to check guest version of Virtualbox Guest Additions on this platform. Stopping installation.
==> default: Checking for guest additions in VM...
default: The guest additions on this VM do not match the installed version of
default: VirtualBox! In most cases this is fine, but in rare cases it can
default: prevent things such as shared folders from working properly. If you see
default: shared folder errors, please make sure the guest additions within the
default: virtual machine match the version of VirtualBox you have installed on
default: your host and reload your VM.
default:
default: Guest Additions Version: 4.2.16
default: VirtualBox Version: 4.3
==> default: Mounting shared folders...
default: /vagrant => /Users/joel/Code/anybots/operations/robot/freebsd92-i386
Vagrant attempted to execute the capability 'mount_virtualbox_shared_folder'
on the detect guest OS 'freebsd', but the guest doesn't
support that capability. This capability is required for your
configuration of Vagrant. Please either reconfigure Vagrant to
avoid this capability or fix the issue by creating the capability.
Note that I have recently installed the latest version of VirtualBox, but somehow I can't find the Guest Additions.
Hi. When I synchronize two folders using Unison, only one of the roots seems to be affected. Below are all the information I would think is necessary to figure out why it is working like it is.
I'm using
$ unison -version
unison version 2.27.57
From the Ubuntu repositories.
My work laptop:
$ echo $UNISONLOCALHOSTNAME
worklaptop
$ pwd
/home/userfoo
$ ls -lAR .unison*
.unison:
total 8
drwxr-xr-x 2 userfoo userfoo 4096 2010-04-26 11:39 backups
-rw-r--r-- 1 userfoo userfoo 231 2010-04-26 11:38 default.prf
.unison/backups:
total 0
.unisonroot:
total 0
$ cat .unison/default.prf
# Roots of the synchronization
root = /home/userfoo/.unisonroot
root = ssh://devel//home/userbar/.unisonroot
path = *
backuplocation = central
backupdir = /home/.unison/backups
backupprefix = $VERSION.bak
$ mkdir .unisonroot/aDirectoryFrom-$UNISONLOCALHOSTNAME
$ echo something >.unisonroot/aFileFrom-$UNISONLOCALHOSTNAME
$ ls .unisonroot/
aDirectoryFrom-worklaptop aFileFrom-worklaptop
And the Ubuntu server I want to synchronize with:
$ echo $UNISONLOCALHOSTNAME
workcmpuserbardevel
$ pwd
/home/userbar
$ ls -lAR .unison*
.unison:
total 4
drwxr-xr-x 2 userbar userbar 4096 2010-04-26 11:38 .unison
.unison/.unison:
total 0
.unisonroot:
total 0
$ mkdir .unisonroot/aDirectoryFrom-$UNISONLOCALHOSTNAME
$ echo something >.unisonroot/aFileFrom-$UNISONLOCALHOSTNAME
$ ls .unisonroot/
aDirectoryFrom-workcmpuserbardevel aFileFrom-workcmpuserbardevel
I perform the unison synchronization:
$ echo $UNISONLOCALHOSTNAME
worklaptop
$ unison
Contacting server...
Connected [//worklaptop//home/userfoo/.unisonroot -> //workcmpuserbardevel//home/userbar/.unisonroot]
Looking for changes
Warning: No archive files were found for these roots, whose canonical names are:
/home/userfoo/.unisonroot
//workcmpuserbardevel//home/userbar/.unisonroot
This can happen either
because this is the first time you have synchronized these roots,
or because you have upgraded Unison to a new version with a different
archive format.
Update detection may take a while on this run if the replicas are
large.
Unison will assume that the 'last synchronized state' of both replicas
was completely empty. This means that any files that are different
will be reported as conflicts, and any files that exist only on one
replica will be judged as new and propagated to the other replica.
If the two replicas are identical, then no changes will be reported.
If you see this message repeatedly, it may be because one of your machines
is getting its address from DHCP, which is causing its host name to change
between synchronizations. See the documentation for the UNISONLOCALHOSTNAME
environment variable for advice on how to correct this.
Donations to the Unison project are gratefully accepted:
http://www.cis.upenn.edu/~bcpierce/unison
Press return to continue.[<spc>] Waiting for changes from server
Reconciling changes
local workcmps...
dir ----> aDirectoryFrom-worklaptop [f]
file ----> aFileFrom-worklaptop [f]
Proceed with propagating updates? [] y
Propagating updates
UNISON 2.27.57 started propagating changes at 11:49:14 on 26 Apr 2010
[BGN] Copying aDirectoryFrom-worklaptop from /home/userfoo/.unisonroot to //workcmpuserbardevel//home/userbar/.unisonroot
[BGN] Copying aFileFrom-worklaptop from /home/userfoo/.unisonroot to //workcmpuserbardevel//home/userbar/.unisonroot
[END] Copying aDirectoryFrom-worklaptop
[END] Copying aFileFrom-worklaptop
UNISON 2.27.57 finished propagating changes at 11:49:14 on 26 Apr 2010
Saving synchronizer state
Synchronization complete (2 items transferred, 0 skipped, 0 failures)
And then check the .unisonroot directory on the computer I started the synchronization from:
$ ls .unisonroot/
aDirectoryFrom-worklaptop aFileFrom-worklaptop
And on the server:
$ echo $UNISONLOCALHOSTNAME
workcmpuserbardevel
$ ls .unisonroot/
aDirectoryFrom-worklaptop aFileFrom-worklaptop
aDirectoryFrom-workcmpuserbardevel aFileFrom-workcmpuserbardevel
As can be seen above, the contents of the laptop .unisonroot has not changed while the servers .unisonroot has. The desired result would have been that the two folders would have ended up being identical, holding the union of the contents of the two roots.
Here is the most interesting thing I've noticed since the problems started:
If I go into a control panel/system module (in this case the Resource Monitor) that has a "Check Online" type option, Firefox (my default browser) opens right up without a problem.
But if I just start Firefox from any shortcuts (start menu, desktop, etc), the Firefox process starts up (and the start menu icon starts glowing) only to end without notice a few seconds later.
Possibly related: If I start up in Safe-Mode (w/o Networking, but haven't tried with yet), I can start up FF or Chrome just fine, but if I attempt to open Chrome normally, I get a permissions error.
Opera and Safari seem to be okay (mostly). Safari crashes when I try to download any files.
All of the above leads me to believe that some (but clearly not all) core files have messed up permissions. Or rather, that I no longer have permission. System still does, based on Firefox opening without fail when the system initiates it.
I've run MS Forefront once in normal mode, Malwarebytes twice in normal mode and once in safe-mode. One trojan found and deleted, but the problem persists.
Two other things worth mentioning:
I accidentally duplicated my library...
I thought I'd try to add the "Internet" folder to my start menu, next to music and downloads.
The first advanced thing I tried was "create new library". I clearly misunderstood what this means. I thought it was a way to add virtual folders to the library (which I thought, in turn, would allow me to choose it as a link on the start menu), but instead it recreated my already existing user folder, AppData and all. I didn't notice this until today.
Then I tried setting permissions for my User folder to full control, recursively...
Confused but not giving up,I thought I could maybe create a shortcut to the NetHood folder manually, but instead got hit with an access denied error. So I tried to change the permission levels for all sub-folders to my user folder so that I had full control. I got several access denied errors along the way.
At this point I gave up, went out, ended up caught in the rain and stuck on a friend's couch and showing up late for work the next day. Thanks for nothing, Microsoft.
When I finally got home today (20 hours later), I noticed that Firefox was acting really strange. I tried opening Chrome to see if the problem was client side or server side, and instead got the above-mentioned "you don't have permission to open this program" alert.
And I think that's the whole story.
Oh, I also did a system restore, but not chose a point from this morning (an auto update), and it worked but the problem wasn't fixed. And then all the earlier restore points were gone.
So the questions are:
a) is there a way to set the admin and user privs back to "default"?
b) would this, in anyone's expert opinion, fix the problems I'm having?
c) how come being logged in as an admin isn't the same as being logged in with admin privs? It seems that half the time I have to do run as admin for fairy standard things because i'm being treated as me-theuser and not me-theadmin.
Thanks for reading.
Hello,
Here's the situation: I went on vacation for a couple of weeks, but before I left, I took the harddrive out of my computer and hid it in a different location. Upon coming back on Monday and putting the harddrive back in my computer, I right-clicked on different files to see their properties. Interestingly enough, several files had been accessed during the time I was gone! I right-clicked different files in various locations on the harddrive, and all of these suspect files had been accessed within a certain time range (Sunday, ?January ?09, ?2011, approximately ??between 6:52:16 PM - 7:16:25 PM). Some of them had been accessed at the exact same time--down to the very second. This makes me think that someone must have done a search on my harddrive for certain types of files and then copied all those files to some other medium. The Windows 7 installation on this harddrive is password protected, but NOT encrypted, so they could have easily put the harddrive into an enclosure/toaster to access it from a different computer.
Of course I did not right-click every single file on my computer, but did so in different folders. For instance, one of the folders I went through has different types of files: .mp3, ,prproj, .3gp, .mpg, .wmv, .xmp, .txt with file-sizes ranging from 2 KB to 29.7 MB (there is also a sub-folder in this folder which contains only .jpg files); however, of all these different types of files in this folder and its subfolder, all of them had been accessed (including the .jpg files from the sub-folder) EXCEPT the .mp3 files (if it makes any difference, the .mp3 files in this folder range in size from 187 KB to 4881 KB). Additionally, this sub-folder which contained only .jpg files (48 .jpg files to be exact) was not accessed during this time--only the .jpg files within it were accessed-- (between 6:57:03 PM - 6:57:08 PM).
I thought that perhaps this was some kind of Windows glitch that was displaying the wrong access date, but then I looked at the "date created" and "date modified" for all of these files in question, and their created/modified dates and times were spot on correct.
My first thought was that someone put the harddrive into an enclosure/toaster and viewed the files; but then I realized that this was impossible because several of the files had been accessed at the same exact time down to the second. So this made me think that the only other way the "date accessed" could have changed would have been if someone copied the files.
Is there any chance at all whatsoever that this is some kind of Windows glitch or something, or is it a fact that someone was indeed accessing my files (and if someone was accessing my files, am I right about the files in question having been copied)? Is there any other possibility for what could have happened?
Do I need to use any kinds of forensics tools to further investigate this matter (and if so, which tools), or is there any other way in which I can be certain of what took place in that timeframe the day before I got back? Or is what I see with Windows 7 good enough (i.e. accurate and truthful)?
Thanks in advance, and please let me know if any other details are required on my part.