Search Results

Search found 14439 results on 578 pages for 'folder customization'.

Page 416/578 | < Previous Page | 412 413 414 415 416 417 418 419 420 421 422 423  | Next Page >

  • word 2010 Macro to name and Save file when opened

    - by Phillip Clark
    I have a word document template and will be using a hyper link in Excel to access the word file. The issue I need to resolve is making sure once it is opened a message field box is "popped" up asking the user to create a new file name ... (in this case the current date) for each time the file is opened. In the message pop when finished entering in file name they click yes and then the save screen comes up with the path/file type (macro enabled document) and the file name they have already entered in the pop up.. All they should have to do from the save screen is click ok and it saves the file to a certain path/folder on the C drive of the computer. Once they finish typing in their notes they click a active x button to save and close and they are finished. If anyone can help with this it would be fantastic.

    Read the article

  • MS Excel Vba/Macro equivalent in LibreCalc or OpenOfficeCalc

    - by ReggieCL
    is there an equivalent macro/vba in libre calc that does this routine; - Read/open xls files in a path and do a batch import/copy of read sheets and merge it with the current open workbook. Here's the vba I used in MS Excel. Thanks in advance Sub Consolidate_Sheets() 'Folder Path to read the xlsx files from Path = "F:\WIP2\Below 25\" filename = Dir(Path & "*.xlsx") Do While filename <> "" Workbooks.Open filename:=Path & filename, ReadOnly:=True For Each sheet In ActiveWorkbook.Sheets 'import/copy sheets from to read xlsx files sheet.Copy After:=ThisWorkbook.Sheets(1) Next sheet Workbooks(filename).Close filename = Dir() Loop End Sub

    Read the article

  • How to disable the text shadow in plasma widgets KDE 4.6?

    - by piedro
    I just switched completely to Kubuntu for the sake of the more suitable applications for my purposes and their overall integration in the system. But I'm not very happy with all those desktop effects and tranparent looks everywhere. Some things a matter of taste I guess. Some things just unbearable for me having weak eyes. One of them is that about every light theme seems to be using a text shadow effect on the plasma widgets. A white "spilled milk" effect underneath dark text on light grey or glassy background. Drives my eyes nuts! I can remove this effect on the desktop folder by unselecting "shadow" as text effect. I can't find any way to switch it of in the panel and the plasma widgets. My second, related question is, is there a plasma theme matching the oxygen look of the default desktop in light colors that uses the same colors as the ones chosen in the KDE color settings, - simple, opaque, no effects, gnome-like? Plz help someone! thx, piedro

    Read the article

  • SMTP and IIS maild from my website

    - by RupDog
    Hi, I am using the free version for Google Apps to host my mail. I want to set is so that mail is routed via google apps as this will give me a much better reputation and mail will not go to the spam folder as much. Doe anyone know how this is done? Would I need to setup IIS to do this? The site is writeen in C# ASP.NET. So Could I perhaps just update the web.config file to route mail via google apps?

    Read the article

  • Unable to get to remote samba share

    - by tubaguy50035
    I have a remote VPS that I would like to setup samba on and only allow my IP access to it. I currently have in my smb.conf: [global] netbios name = apollo security = user encrypt passwords = true socket options = TCP_NODELAY printing = bsd log level = 3 log file = /var/log/samba/log/%m debug timestamp = yes max log size = 100 [hosting] path = /hosting/ comment = Hosting Folder browseable = yes read only = yes guest account = yes valid users = nick I have the ports (137,138,139,445) open in iptables (they're open to everyone right now while I debug) and I see nothing in the syslog about iptables blocking my requests. When I try to open a file browser to my address \\ipaddress, it hangs for a good thirty seconds, and then opens a log in box. I enter my user name and password for the server, hit okay. It then opens the same box, I enter my credentials again and hit enter. Windows then tells me it could not connect. My user account is added to Samba already. Anybody have any suggestions what I can do to get this working?

    Read the article

  • Change ownership of directory and all contents to a new user from root.

    - by Andrew Fashion
    I created a website under /var/www/html/ all under root, all images, files, .htacess, directories, etc... I uploaded and configured everything as root. I want to make it it's own username/password so it's not owned by root. I currently do not have the user account made either, I want to also setup FTP for the user account. There is also about 30GB of images in the folder as well. How can I go about changing all of this? I am running CentOS 5.5 64 bit. Thank you!

    Read the article

  • remove apache tar binary package in centos

    - by user119720
    I need help in removing binary package that had been installed in my linux machine. The scenario that I am having are such as like this: I've already install latest apache that I get from its website(httpd.apache.org) through Unix binary package(tar.gz) After successfully install program,the apache server web perfectly without any issues. But then I having a thought,if there are another latest release version of apache in the future,then I need to make sure that i can upgrading current apache or reinstalling the new version apache. So my question is, how do I ensure that I have remove the old remove apache and all its dependencies so that it will not having conflict(probably) when installing the new apache. Right now the only thing that i can think of is to remove all the apache folder manually : rm -rf /apache2 Hope someone can shed some light about this.Thanks.

    Read the article

  • Run external application on markdown source in ikiwiki

    - by student
    Can I add a button to each wiki page in ikiwiki which launches an external application (on the client side) or script with the markdown code of the current page as input? Edit: I didn't realize that it might be complicated to do it on client side as Zenklys' answer suggested. So perhaps I should describe more concretely what I have in mind: I want to have two buttons: "Get LaTeX" and 2. "Get pdf". Clicking on "Get LaTeX should generate a LaTeX file and the browser should simply open or download that file. Analogously for the pdf. It would even be ok, to have a button "Generate LaTeX" instead, which generates the LaTeX code and changes after the generation to "Get LaTeX" which simply points to the LaTeX file. So it is not really necessary to do the generation of the files on client side. Would be ok, if this is done (on a temporary folder) on server side. For the LaTeX resp. pdf generation I want to use a custom wrapper script for pandoc, let's call it mymarkdown2latex resp. mymarkdown2pdf.

    Read the article

  • How can I stream video from my XP sp3 desktop to my win 7 laptop?

    - by Robb
    I would like to stream media from my XP sp3 desktop to my Win7 laptop. I currently have Windows Media Player 11 on both machines. The XP one is set up for the media sharing service as far as I can tell. I have also set the 'Play To' option to my Win 7 machine on the XP WMP. However, the only way I've been able to 'stream' is to simply tell my Win7 WMP to add a network folder to its libraries. This seems to work over the wifi but it does stutter a touch, even with a 20 second buffer. The stuttering also appears randomly, maybe 4-5 times during a 22 minute show. I feel like I'm doing something wrong. Is this setup OK or am I missing something?

    Read the article

  • Importing an existing project into Git

    - by Andy
    Background During the course of developing our site (ASP.NET), we discovered that our existing source control (SourceGear Vault) wasn't working for us. So, we decided to migrate to Git. The translation has been less than smooth though. Our site is broken up into three environments DEV, QA, and PROD. For tho most part, DEV and the source control repo have been in sync with each other. There is one branch in the repo, if a page was going to be moved up to QA then the file was moved manually, same thing with stuff that was ready for PROD. So, our current QA and PROD environments do not correspond to any particular commit in the master branch. Clarification: The QA and PROD branches are not currently, nor have they ever been in source control. The Question How do I move QA and PROD into Git? Should I forget about the history we've maintained up to this point and start over with a new repo? I could start with everything on PROD, then make a branch and pull in everything from QA, and then make another branch off of that with DEV. That way not only will the branches reflect the differences in the environments, they'll be in the right order chronologically with the newest commits in the DEV branch. What I've tried so far I thought about creating a QA branch off of the current master and using robocopy to make the working folder look like the current QA environment. This doesn't work because the new commit from QA will remove new files from DEV and that will remove them when we merge up, I suspect there will be similar problems if I started QA at an earlier (though not exact) commit from DEV.

    Read the article

  • wget recursively with -np option still ascends to parent directory

    - by vectra
    tl;dr: will `wget --no-parrent -r ' download from a directory above the given url's directory? when using wget to download, say images, recursively from example.com/a/b with the -r and -np options, will a picture that is under example.com/a/c/ be downloaded when example.com/a/b/ delivers a html-file containing a link to the picture? if so, how do i get all pictures, that are in a folder and it's subfolders and only those? the description of the option --no-parent says "Do not ever ascend to the parent directory when retrieving recursively". anyway directory browsing delivers a link to the parent directory, which wget will follow, despite mentioned option. now what did i miss? edit: using GNU Wget 1.12

    Read the article

  • How can I force the Pictures app to refresh its library?

    - by Rowland Shaw
    I've imported some pictures to my "My Pictures" folder using Adobe Lightroom, but the Pictures app doesn't seem to notice the new files? Is there some way to force it to refresh its library? To be clear, Pictures are sorted under C:\Users\<username>\Pictures\ by year, then date; for example: C:\Users\<username>\Pictures\2012\2012-10-20, these show, however those in C:\Users\<username>\Pictures\2012\2012-10-29 do not, despite coming from the same camera, being in the same format, thumbnails working in Exporer, the pictures opening in the Pictures app via a double click in explorer, and Windows Media Centre having no problems. I have tried excluding the Pictures library, and reincluding it, but that had no effect. It has taken three or four reboots for even the folders to show up in the app, but it reports them as empty

    Read the article

  • Seemingly random 404's for static files in Pyramid project

    - by seth
    I'm running a Pyramid project with mod_wsgi. Some of the files in my static directory (images, stylesheets, javascript) load fine, but others are coming up as not found. The files that are not working are all web fonts (otf, svg, woff and eot). I tried adding a text file into the static directory where the fonts are to see if I could access it, but it also came back with 404. The same text file also can't be accessed when put in the images folder. From what I'm looking at, it doesn't seem to be a permissions issue. Any ideas?

    Read the article

  • How is a Chromebook better than using Chrome browser on another OS?

    - by bigpopakap
    I'm looking into getting a Chromebook as a lightweight device to use for basic functions. I'm fully aware of some of the limitations like no native apps, limited functionality while offline (though this is getting better with Google Drive's offline capabilities). Currently, I have a Windows machine on which I've installed Google Drive, so files in that folder are synced. And I use Chrome as my browser. So I have access to all the same apps (Drive, Calendar, Gmail, Google Music, etc.). Is there any advantage to having a Chromebook over my current setup, other than the speed of the lightweight Chrome OS?

    Read the article

  • Do you have to recreate workspaces after upgrading a TFS 2008 server to TFS 2010?

    - by Clara Oscura
    I am just reposting this thread from a MSDN forum since it seems to be unavailable. It was very useful when I was having trouble with my folder mappings after migrating to TFS 2010. Question: I opened VS2008 and connected it to the upgraded 2010 TFS server.  Upon clicking any of our Team Projects in source control explorer I get "Team Foundation Error - The workspace MYWORKSPACE;DOMAIN\MYUsername already exists on computer MYPCNAME." Answer: The same local paths on your machine are mapped to 2 different workspaces, one on the preupgrade server and one on the postupgrade server.  It's not safe to have multiple workspaces on different servers mapped to the same local paths b/c you could pend some changes while connected to one server, and the other server would have no idea what you did.  You should either delete your conflicting workspaces from one of the servers (if you don't need them on both), or test the new TFS instance from a new workspace (on different machine). If you want to test an existing production workspace on both servers, then yes, you will have to mess around with the workspace cache. You don’t have to delete the entire cache, you just need to run "tf workspaces /remove:* /server:<serverurl>" to clear the cached workspaces from a server (the command won't delete the workspaces), and possibly "tf workspaces /server:<server>" to refresh the workspace cache for a given server.  You will also have to do back up and restore the workspace before switching servers or your local files could be inconsistent. From the “Microsoft Visual Studio Team Foundation Server 2010 Beta 1” forum (not available anymore?) Technorati Tags: TFS 2010,TFS Workspaces,Team System,Team Foundation Server 2010

    Read the article

  • Backing up data (including mysqldumps) to S3

    - by seengee
    We have a web app on a number of servers and we want to add an additional layer of redundancy by backing up the key data to S3. The key data is the MySQL database and a folder containing dynamically created site assets - predominantly images. Some kind of rsync based solution would initially seem the best plan. A couple of years ago we played with S3cmd (in particular s3cmd sync) with some success but we didn't find it particularly reliable although this may have changed since. Its occurred to me though that a rsync solution might not work particularly well with a single db.sql file created with mysqldump and I assume this means the whole database getting transferred each time, with multiple databases of over 1GB this is going to add up to a lot of traffic (and $s) very quickly. With the image files I could simply just transfer files modified within the last day which would be far more simple. What approach should I look at?

    Read the article

  • Run Grunt task in Visual Studio Release Build with a bat file

    - by Aligned
    Originally posted on: http://geekswithblogs.net/Aligned/archive/2014/08/19/run-grunt-task-in-visual-studio-release-build-with-a.aspx 1. Add a BeforeBuild in your csproj file. Edit the xml with a text editor. <Target Name="BeforeBuild"> <Exec Condition="'$(Configuration)' == 'Release'" Command="script-optimize.bat" /> </Target> 2. Create the script-optimize.batREM "%~dp0" maps to the directory where this file exists cd %~dp0\..\YourProjectFolder call npm uninstall grunt call npm uninstall grunt call npm install --cache-min 604800 -g grunt-cli call npm install --cache-min 604800 grunt typescript requirejs copy less:compile less:mincompileThis grunt command will compile typescript, run the requireJs optimizer, complie and minimize less.3. Make it use the minified code when the Web.config compilation debug is set to false <!-- These CustomCollectFiles actions are used so that the Scripts-Release folder/files are included        when publishing even though they are not project references -->  <Target Name="CustomCollectFiles">    <ItemGroup>      <_CustomFiles Include="Scripts-Release\**\*" />  </ItemGroup>  </Target> That should be all you need to get a Grunt task to minify and combine JS (plus other tasks) in Visual Studio Release build with debug = false. This is a great video of Steve Sanderson talking about SPAs, npm, Knockout, Grunt, Gulp, ect. I highly recommend it.

    Read the article

  • Accidentally deleted the software for MyPassport Essential SE 1TB Hardrive

    - by user26192
    I'm posting for a friend of mine. She bought a WD MyPassport Essential SE 1 TB Hard drive the other day. When she plugged in the USB in her lap top, the driver cannot be recognized by the smart ware software. While she was doing a back up of her files, McAfee was running in the background. Since the backup was taking so long to finish, she decided to pause it. She tried to delete the partially backed up files, but instead, she accidentally deleted the entire file in the folder including the pre-installed software. Now, when she tries to start up the MyPassport, the smart ware doesn't show up anymore. Can someone please give us advice what can she do about this? Thank you.

    Read the article

  • Accessing server by dedicated IP address

    - by Sherwin Flight
    I'm having an issue with my hosting provider after migrating to a new account. It's taking some time to get the problem sorted out, so I am hoping someone here can shed some light on the situation. The server is running WHM/cPanel, and the site I am trying to access has a dedicated IP address. When I connect to the server like this http://IP.HERE instead of showing my the website the way I would expect, it is showing the contents of a subfolder. So, while I would expect it to load public_html/ it is loading public_html/somefolder/ instead. Any idea why this is happening instead of showing the sites homepage the way I would expect? EDIT It is not redirecting, so the url is just http://IP.ADDRESS/, but the files listed are from a subfolder. So, it LOOKS at though I went to http://IP.ADDRESS/subfolder, when the URL says it should be showing the main folder contents. When I access the site using the domain name, it works properly, so I assume the document root is set correctly.

    Read the article

  • Parent who hates change [closed]

    - by Ian Boyd
    My mother was nearly brought to tears because My Pictures is no longer under My Documents. Everything in Windows (7) works around My Pictures being in its own location, with "Libraries" and everything. Her Picasa wants to index My Pictures, but finds it empty. What's more is that Microsoft created, as a compatibility hack, a junction for anyone who hard-codes My Pictures as being a sub-folder of My Documents: <JUNCTION> My Pictures [C:\Users\Mom\Pictures] What's more is that there is Pictures ? My Pictures I'm about ready to tell her that her new computer is being taken away until she either accepts the changes, or dies. Suggestions?

    Read the article

  • vim + Ruby on Rails: how do you bounce among those 4-5 files you're currently working on?

    - by glitch
    I'm just starting to get familiar with vim, and I'd like to use it as my primary Rails development tool. As a Visual Studio and RubyMine user, I find a lot of stuff to be missing from the barebones vim installation, and therefore I went ahead and attempted to soup it up with plugins such as: rails.vim tcomment ruby-vim NERDtree and a couple of others. The issue is that I still don't quite get the average work-flow of using vim as one's Rails IDE. In RubyMine (again, similarly to Visual Studio) I have a series of tabs always open, containing the main files I'm switching among, and I additionally use NERDtree to open files from the folder structure. I tried opening them as new tabs, but the tab system in vim is just a lot more awkward than that in real IDEs. (I haven't seen vim pros in action, but I imagine that they'd not be relying on tabs, but using numerous splits instead, keeping at least a couple of files per split and switching between them with CTRL + ^. Is that the case?) So, at the end of the day, how do I really squeeze the most from vim if I want to be able to quickly access several files at once? Thank you!

    Read the article

  • Using a CDN for CMS software (multiple sites)

    - by SmokeyPHP
    I'm currently researching ideas for the media management side of a CMS I'm writing. I was looking at having images served from a CDN which is fine on a single site, but I want all sites that run the CMS to make use of a CDN (which will most likely be a custom developed one, rather than a third party service like S3). My main question is: Is a multi-site CDN a good idea? I can't think of a downside, but have probably missed something - obviously they won't share the same folder, as I invisage the requests to be css.cdnsite.com/example.com/style.css or something along those lines. Having multiple sites in the same place will obviously make it easier for us to manage, as well as being cheaper, but then I wonder if it'll be worth it... Long story short: How should the CMS handle user uploaded media (separate installations) Just keep a local copy of all assets and serve them from the same site, like in days of yore? Keep a local copy, force site to use www. and have CDN subdomains per site? Or use a single separate CDN for all sites? Apologies for the length of this question, not sure if this should be multiple questions or not, as all parts are kind of related and could affect each other.

    Read the article

  • Did bittorrent sync take advantage of multiple peers

    - by NachoChip
    I am currently testing BT sync, and really like it. However, I cannot find much of the technical detail regarding to how it works. I wonder did anyone know that is the program take advantage of multiple peers connection? For example, if I have a File A(FA), which exist in Computer 1 (C1) and Computer 2 (C2). If I set up a sync from C1 and C2 to a folder in Computer 3 (C3), will I get a better sync up time because I can get FA from 2 sources at the same time (of course, assume I have enough downstream)?

    Read the article

  • Cannot install VS Team System 2008 on Windows 7 Ultimate (64bit)

    - by systemX
    Hello, i am trying to install VS TS 2008 on W7 Ultimate (64bit), but i have run into errors during the setup. Please take note that i have tried to mount the iso to a virtual drive, and also extracted the iso contents to a local folder. Both methods have failed and produce the same error log below. [10/26/09,03:02:40] Runtime Pre-requisites: [2] Error: Installation failed for component Runtime Pre-requisites. MSI returned error code 1603 [10/26/09,03:02:42] VS70pgui: [2] DepCheck indicates Runtime Pre-requisites is not installed. [10/26/09,03:02:42] VS70pgui: [2] DepCheck indicates Microsoft Visual Studio 2008 64bit Prerequisites (x64) was not attempted to be installed. And the list goes on and on.. This is a fresh install of W7, and i have not installed MS Office 2007 at all yet, not sure if it would be causing my errors right now.. I appreciate any help i can get thank you.

    Read the article

  • Can I fine-tune my screen brightness? (Ubuntu Studio, Thinkpad T420s)

    - by Dan S
    I am in a similar situation as question Can I fine-tune my screen brightness? . For me, Thinkpad T420s and Ubuntu Studio 11.10. The brightness keys work fine but the increments are too large. I have the /sys/class/backlight/acpi_video0 and I can see the value of actual_brightness changing as I use the brightness keys - it changes between values 0,3,6,9,12,15. None of the listed answers for fine control works here, as follows: As the root user I can echo integers to brightness in the above folder. As an ordinary user I cannot. I do not have /usr/sbin/gnome-power-backlight-helper (because I'm using xfce perhaps?), and not sure what package it's from or if installing such a package would make sense since it's not what's currently managing the backlight. How can it be done, please?

    Read the article

< Previous Page | 412 413 414 415 416 417 418 419 420 421 422 423  | Next Page >