Search Results

Search found 21454 results on 859 pages for 'via'.

Page 124/859 | < Previous Page | 120 121 122 123 124 125 126 127 128 129 130 131  | Next Page >

  • Same album repeatedly updates when syncing iTunes with iPhone

    - by Unsliced
    I have many many albums and tracks purchased via the iTunes store both through iTunes directly and via the iPhone. They all synchronise properly and are on iTunes eventually (i.e. if purcashed originally on the iPhone). There's one album, though, purchased via the iTunes app in June this year. Virtually every time I sync the iPhone it says "updating files" and this album gets updated. The rest of the sync goes just fine, new tracks added, listened to podcasts are removed, playlists are updated (notwithstanding that smart playlists' orderings are now utterly broken) fine - but this album is always getting updated. In finder I've looked at the permissions and nothing seems unusual. FWIW, I'm on Mac OS X 10.5.8, iPhone 3.1.2, iTunes 9.02. Anyone with any ideas? It's not a fundamental problem, but it does means that each sync takes an extra 30-60 seconds, which is annoying!

    Read the article

  • Symfony sfControlPanel 500 error only when executing 'tasks'

    - by rrlange
    As a bit of background, I recently had to restore a Symfony site from backup. Ever since the 'successful' restore, I am running into an exception when trying to execute a(ny) task via the web-based sfControlPanel: Unable to find PHP executable stack trace * at sfToolkit::getPhpCli() in SF_ROOT_DIR/plugins/sfControlPanelPlugin/modules/sfControlPanel/actions/actions.class.php line 93 FYI: symfony version 1.0.6 PHP 5.2.6 (cli) (built: May 2 2008 16:06:40) Apache/2.2.3 CentOS 5.* Thank you very much for any and all suggestions as to what might be amiss. Addendum: I neglected to mention that I can the frontend app (and certain backend apps) perfectly fine via the web. I am also able to run common tasks via the command line (cc etc.).

    Read the article

  • FREE three days of online SharePoint 2010 development training for UK software houses Feb 9th to 11th

    - by Eric Nelson
    I have been working to get a SharePoint development course delivered online in February and March – online means lots of opportunities to ask questions. The first dates are now in place. The training is being delivered as a benefit for companies signed up to Microsoft Platform Ready. It is intended for UK based companies who develop software products* Agenda: Day 1 (Live Meeting 3 hours) 1:30 - 4:30 •         Getting Started with SharePoint: Understand why and how to start developing for SharePoint 2010 •         SharePoint 2010 Developer Roadmap:  Explore the new capabilities and features •         UI Enhancements: How to take advantage of the many UI enhancements including the fluent UI ribbon and  extensible dialog system. Day 2 (Live Meeting 3 hours) 1:30 - 4:30 •         Visual Studio 2010 Tools for SharePoint 2010: Overview of the project and item templates and a walkthrough of the designers •         Sandboxed Solutions: The new deployment model can help mitigate the risk of deploying custom code   •         LINQ to SharePoint:  SharePoint now fully supports LINQ for querying lists Day 3 (Live Meeting 3 hours) 1:30 - 4:30 •         Client Object Model: The Client OM can be accessed via web services, via a client (JavaScript) API, and via REST •         Accessing External Data: Business Connectivity Services (BCS) enables integration with back end systems •         Workflow: A powerful mechanism to create functionality using Windows Workflow Foundation Register for FREE (and tell your colleagues – we have a pretty decent capacity) To take advantage of this you need to: Sign your company up to Microsoft Platform Ready and record your SharePoint interest against one of your companies products Read about Microsoft Platform Ready Navigate to the “Get Technical Benefits” tab for SharePoint and click on Register Today You will then ultimately get an email with details of the Live Meeting to join on the 9th. But you should also favourite the team blog for any last minute details * Such companies are often referred to as an Independent Software Vendors. My team is focused on companies that create products used by many other companies or individuals. That could be a packaged product you can buy "off the shelf" or a Web Site offering a service - the definition is actually pretty wide these days :-) What it does not include is a company building software which will only be used by its own people.

    Read the article

  • Netgear genie says my Internet is off, while Windows claims I am connected

    - by Manu
    I'm connected via wifi to my ISP's router/modem. While Windows says that I'm always connected, I keep getting messages from Netgear Genie that I've lost contact to the Internet, and I cannot access webpages until it comes back. There are two other computers in the house, one connected to the router via ethernet, the other via wifi, both seem to have no such problems. I've wondered if Netgear genie itself was the problem, but I am regularly disconnected even if I uninstall it. And I'd rather have it since it accurently tells me if I'm connected or not. Why does windows says I'm online if I can't access any online game, or website ? Is "netgear genie" the problem ? I've removed the connection, recreated it, I've even copied the settings to a usb key from the computer that has wifi access.

    Read the article

  • Route eth0 to internet traffic and eth1 to local traffic

    - by Romain Caire
    How can I route all my internet traffic on eth0 (everything except 192.168.1.0/24) and route my local traffic through eth1 (192.168.1.0)? Here is my attempt : # Flush ALL THE THINGS. ip route flush table main # Restore the main table. I flushed it because OpenVPN does weird things to it. ip route add 127.0.0.0/8 via 127.0.0.1 dev lo ip route add 0.0.0.0/0 via 164.67.195.1 ip route add 192.168.1.0/24 via 192.168.1.1 ip route flush cache

    Read the article

  • Best Practices in Preventing Locks in File sharing in Windows?

    - by crosenblum
    We have a web development server, that has our source control on it. That we use via mapped drives to open/edit/save/create files on to it. But I have noticed a lot of weird glitches with either the file lock/sharing/etc For example, if i have a file open on my local pc, and try to check out that file via my source control on dev, it error's out. The source control I use is QCVS by Qumasoft. Or If I use cuteftp to download something to dev, via my local pc, it won't download unless I close that file. So my brain sniff's and figures this is either an issue with folder/network/sharing, or an issue of permissions. Has anyone else had similar issues, and what did they do to fix it permanently? Thank you...

    Read the article

  • Missing menu items for Azure SQL tables within SQL Server Management Studio?

    - by Sid
    I have a table (say Table1) that is replicated via SQL Data Sync Agent across a local SQL Server 2012 as well as an Azure SQL Server (part of Microsoft Azure). Everything about Table1 (schema, table values etc ) is identical to the best of my understanding. However, when I list and right click Table1 from Microsoft SQL Server Management Studio 2012 (SSMS), I get some very different menu options, even for seemingly basic stuff. Lets focus only on the 'Design' menu item: It is visible for Table1 on the local SQL server in SSMS It is missing for Table1 on Azure SQL via SSMS It is visible for Table1 (as Open Table Definition) on Azure SQL when reaching it via Visual Studio 2012 (Server Explorer - Data connections) This is seen in the screenshots below: Now I use scripts from some real stuff (esp when I need to check in the SQL scripts etc) but this difference concerns me to some extent. Am I witnessing just a tools artifact in SQL Server Management Studio when connecting to Azure SQL? or is it something more serious about limitations of Azure SQL itself (although, just seeing the Design surface is so basic!)?

    Read the article

  • Streaming audio to headless linux box

    - by Ralph
    I have a dual boot (Win 7 + Ubuntu) PC connected via wifi with my music collection on a local HDD. I usually use Rhythmbox on Ubuntu or Winamp on Windows to listen to my music but I'll change if I have to. I also have a Raspberry Pi (low power PC running Debian) in the living room that is usually headless and connected via ethernet. The Raspberry Pi is also connected to my living room speakers via an amp. I would like to be able to stream music from my PC over the network to the linux raspberry pi. What software can I use to do this? Some sort of audio client\server?

    Read the article

  • SSH: one user logs in, other don't

    - by Co Lega
    Some users can ssh correctly, others don't. I have an admin user, which is in wheel. I used vsftpd for allowing FTP. I created a nologin user for FTP. Call this user "ftpuser". Now I want to allow sftp. It allows me to connect using admin user. I remove the nologin, by giving the user a shell via usermod -s /bin/bash ftpuser. It still doesn't allow me to connect from the ftpuser via sftp. The content of /home/ftpuser/.ssh is just the known_hosts file which contains "localhost" entry only. User permissions are (in theory) ok: ls -la /home working (admin) : drwx------ 18 admin admin 4096 Feb 6 15:33 admin non-working(ftpuser): drwx------ 3 ftpuser ftp 4096 Mar 26 15:25 ftpuser I haven't configured anything special on openssh. Does the ftpuser need anything extra than shell to enter via ssh?

    Read the article

  • Wireless only connects to Google

    - by dmanxiii
    I am using a netgear wireless router to connect to create a wireless network. When connected to the router via an Ethernet cable everything works correctly. When connecting to the router via wireless only Google can be accessed. I've done searches via Google using random words to make sure that it is not a cached page, and these all work. However, trying to connect to any non Google domain the page fails to load. I've accessed the connection with two different laptops, both are running Windows Vista. The network connection is strong on both, but the issue persists for both. Edit: I have configured the wireless networks name and password, and connect to that using the credentials I have defined. I am sure I am not connecting to a neighbors wireless network. Edit2: Typing in the direct IP address for CNN.com ( http://157.166.255.19 ) would not resolve.

    Read the article

  • How to increase Remote Desktop resolution?

    - by BugAlert
    I'm using Windows 7 on a PC and connect to it via Remote Desktop Connection for Mac by MS, using a Mac. On the Mac, I have a 27" Display with 2560 x 1440 px resolution. The PC has a high performance graphics card (one of those that needs an additional power supply). Both machines are connected via network cable to a router. So network should be insanely fast. However, the max resolution I can get is 1400 x 1050. The PC just has a old and bad 21" monitor connected to it. How can I increase the resolution to 2560 x 1440 to work fullscreen with the Apple LED Cinema Display via Remote Desktop? Maybe I shouldn't even ask, because I already experience some lag at the low resolution of 1400 x 1050. But I guess this can be resolved as well?

    Read the article

  • How to test Laptop NIC's throughput using a router and PC - without be bounded?

    - by 0x90
    My setup includes: Cisco router An i-7 PC running windows A laptop with high speed wifi nic, which I want to check its throughput. I would like to run an FTP server on the PC. hook the router over cables to the PC. I would like to have the PC create its own subnet accessible via the cisco router that would be hooked directly to the PC's nic. From the laptop I want to connect via wifi to the PC's wireless router and connect to the ftp server on the PC. is it possible? how do i connect the router to the PC nic and make it broadcast a subnet via wifi for my laptop to connect to? how do i configure an FTP server to operate only on this subnet?

    Read the article

  • Windows 7 Natively on Mac OS X Bootcamp, Airport Wifi Unable To Connect

    - by Goober
    I am using a brand new MacBook Pro. I am running a copy of Windows 7 natively via bootcamp (No use of Virtual Machine Software at all). However the only way I can get Windows to connect to the internet is via ethernet, as opposed to the Macs Airport card picking up the wireless. It just refuses to connect, and gives me a limited access status. Any ideas!? I've run Windows XP natively via bootcamp and I had a few issues with the network constantly dropping out, however I blamed that on the drivers and the general shiteness of XP.... Help greatly appreciated.

    Read the article

  • Ubuntu server is dropping SSH connections, then not allowing me to log back on

    - by wilhil
    I have an ESX box which I have loaded with two Ubuntu Server machines. During setup, I chose no additional packages to install as I just wanted a lightweight machine for testing. The first thing I did was change the root password via sudo passwd After ESX got on my nerves through lag, I decided to install OpenSSH via apt-get install openssh-server. It did it's business, and I then opened putty and could connect in to both machines fine. The first time it connected, it asked me to add the ssh key as obviously it did not know it. Anyway, the second server is working flawlessly, but, the first seems to be giving me trouble. I was in the middle of typing a sentence when it kicked me off for no reason and when I tried to reconnect, putty gave me a warning that the ssh key had changed and it is potentially dangerous. I attempted to log in anyway and it did not work, just the standard access denied message. Using the second machine, I SSHed in to the first machine and it worked straight away, I then killed the SSH sessions (and possibly SSH server), I then reconnected via putty and I again received the security warning message, but, it allowed me to log on fine. ... I thought "glitch" and nothing more of it, but, it just happened again! I really do not understand this and was hoping someone here can help?

    Read the article

  • Update a DNS to a for a dynamic IP

    - by zobgib
    I want to use my schools connection as a place to host a small webserver but one problem I have run into is anytime my server reboots I am given a new IP inside the schools range. All of the schools IP are public and therefor I can access my computer directly over WAN just via the IP given in ifconfig. I would like to be able to give my computer a dns which is easy enough when I change the Arecords to match the current IP of my computer. The problem is if my computer ever reboots (my school regularly cycles power at night and over holidays) I am assigned a new IP and have to realize it then update the Arecords This is inconvenient and I figure there must be a better way to keep the DNS records updated either via a script or my own BIND server. That way if there is a power cycle I can still access the server via a Domain Name. If you have any direction to point me in it would be much appreciated. I am running Ubuntu 10.04 if that helps :).

    Read the article

  • connecting internet TV modem to repeater router, will it work?

    - by Sandro Dzneladze
    I've internet TV at home, it works via special modem which connects to router via Lan interface. I'd like to move Tv to a room which has no router. so i'd like to use wifi for internet TV. My plan is this, buy another wifi router, set it to repeat sygnal of primary router and attach this TV modem to repeater router via Lan interface. Will this work? I have limited understanding of how internet TV works, so I'm not sure if my strategy will work... does router have to have some special feature to allow this service? will my strategy work?

    Read the article

  • ftp users configuration in OpenSuse 12

    - by chieroz
    I usually work with MacOSX servers but this time I need to set up a ftp service on a OpenSuse 12.2 server and I am a little lost. I am using the remote YAST2 tool via ssh. I created several users who can connect via ssh and/or ftp, so the basic setup is ok. But when connecting via ftp all my users don't have write permissions. The FTP directory for authenticated users is /srv/www/htdocs, which has permissions root:root. The OpenSuse manual say it's bad practice to change these permissions, but my normal users (even the ones in the sudoers list) cannot upload files. So I am stuck: as a workaround I use rsync, but from time to time I just need to establish a working ftp connection. What's the right approach for users permissions in this scenario? Thanks a lot.

    Read the article

  • Keep IMAP messages locally when deleted remotely

    - by user74328
    I use my email from my phone and my computer via IMAP. I want to set something up so that if I delete a message via my phone, my computer will still keep the message locally. For example, assume I leave my computer on, with a synchronize interval of 5 minutes. I want to be able to send something from my phone, wait 5 minutes to be sure my computer has downloaded the item from the Sent folder, then delete it from the IMAP sent folder via my phone, but have the computer at home keep it's copy. Is this possible with any readily available email clients out there? I have Thunderbird and Outlook at the moment, but would be willing to learn a new interface for this feature. How can I accomplish this?

    Read the article

  • Mount remote drive by port forwarding

    - by Anushka
    I have a windows 7 computer at home and a remote windows XP machine with an external USB drive attached to it. I would like to be able to copy files to and from home/the remote USB drive. The problem is that the only way of contacting the work computer is by first sshing into a different remote linux machine as there is a firewall that prevents direct internet connection. I can see two possible solutions but I don't know how to do them. Set up winscp using ssh port forwarding via the linux box so that I can copy to and from the two windows machines. I assume that I would need to run an ssh server on the XP machine as well to do this? Mounting the remote USB drive on the remote windows machine from the home XP machine via ssh port forwarding again via the linux box. Does anyone know how to do either solution?

    Read the article

  • Inbox should not contain any calendar appointment

    - by cotablise
    Imagine two users A and B.User B has editor rights on mailbox A via delegation. Delegation settings: User A should not receive any invitation (calendar appointments). All should be routed to user B. It works fine when you are connected to mailbox via Outlook. Problem is that sometimes it does not work when user is connected via his iPhone (he receives calendar appointments in inbox). I know that this feature for MS products and therefore it works on MS Outlook. But user A told me that it worked in the past also on iPhone with user C. Problem is that nowadays it does not work perfectly for user B. Do you have any suggestion please? Thank you in advance.

    Read the article

  • is it possible to use multiple proxies for different target websites?

    - by fduff
    Typically I'd like to know if it's possible to have a setup where it maps some IPs with a particular proxy and other IPs using another proxy? Proxy Chaining isn't really what I'm looking for (well I don't think it is) because each address could be resolved with or without a proxy but their content may not be accessible without the right proxy. This setup could be part of home network, so that the relevant proxy s/w could be installed and setup on a Linux box for ex. Ex: website-A via proxy-1 website-B via proxy-2 all others via no-proxy

    Read the article

  • How to get around OS X Lion Server SMB "safe save"?

    - by borrrden
    First, there is this problem: http://support.apple.com/kb/TS4149 Which says that anything copied to os x server via smb will be "safe saved" and essentially chmod to 600. This is extremely annoying and counterproductive to having a share server... I did what the article said, but nothing has changed. It is very vague about how to actually set the ACL permissions, but I think I did that via Server.app - hardware - storage - edit permissions. I added (in addition to other things): Guests - Read & Write Others - Read & Write Actually every entry is read and write, but despite this...it is still chmod to 600 when I try to copy a file via SMB. Where else should I look to solve this madness? Alternatively, how can I get rid of this crappy SMB and put back the original SMB? I tried SMBUp but it just fails to start the service every time...

    Read the article

  • Recover snap server data

    - by Ugg
    Hi I have a snap server 110 the machine powers on ok and the healthcheck passes but unable to connect no responce on the assigned ip or any ability to reach the device via the snap server manager. Believe the device is powering on but not loading the OS. Tried pulling the disk running and hooking up to a windows PC via USB, and using disk internals linux reader I am unable to access two of the partitions. ( one of which is the large data partition). There are three partitions on the the drice only one is accessible via Linux reader. I am looking to recover the data of the drive can anyone suggest a DIY option please?

    Read the article

  • web application with secured sections, sessions and related trouble

    - by spirytus
    I would like to create web application with admin/checkout sections being secured. Assuming I have SSL set up for subdomain.mydomain.com I would like to make sure that all that top-secret stuff ;) like checkout pages and admin section is transferred securely. Would it be ok to structure my application as below? subdomain.mydomain.com adminSectionFolder adminPage1.php adminPage2.php checkoutPagesFolder checkoutPage1.php checkoutPage2.php checkoutPage3.php homepage.php loginPage.php someOtherPage.php someNonSecureFolder nonSecurePage1.php nonSecurePage2.php nonSecurePage3.php imagesFolder image1.jpg image2.jpg image3.jpg Users would access my web application via http as there is no need for SSL for homepage and similar. Checkout/admin pages would have to be accessed via https though (that I would ensure via .htaccess redirects). I would also like to have login form on every page of the site, including non-secure pages. Now my questions are: if I have form on non-secure page e.g http://subdomain.mydomain.com/homepage.php and that form sends data to http://subdomain.mydomain.com/loginPage.php, is data being send encrypted as if it were sent from https://subdomain.mydomain.com/homepage.php? I do realize users will not see padlock, but browser still should encrypt it, is it right? If on secure page loginPage.php (or any other accessed via https for that instance) I created session, session ID would be assigned, and in case of my web app. something like username of the logged in user. Would I be able to access these session variable from http://subdomain.mydomain.com/homepage.php to for example display greeting message? If session ID is stored in cookies then it would be trouble I assume, but could someone clarify how it should be done? It seems important to have username and password send over SSL. Related to above question I think.. would it actually make any sense to have login secured via SSL so usenrame/password would be transferred securely, and then session ID being transferred with no SSL? I mean wouldnt it be the same really if someone caught username and password being transferred, or caught session ID? Please let me know if I make sense here cause it feels like I'm missing something important. EDIT: I came up with idea but again please let me know if that would work. Having above, so assuming that sharing session between http and https is as secure as login in user via plain http (not https), I guess on all non secure pages, like homepage etc. I could check if user is already logged in, and if so from php redirect to https version of same page. So user fills in login form from homepage.php, over ssl details are send to backend so probably https://.../homepage.php. Trying to access http://.../someOtherPage.php script would always check if session is created and if so redirect user to https version of this page so https://.../someOtherPage.php. Would that work? 4.To avoid browser popping message "this page contains non secure items..." my links to css, images and all assets, e.g. in case of http://subdomain.mydomain.com/checkoutPage1.php should be absolute so "/images/image1.jpg" or relative so "../images/image1.jpg"? I guess one of those would have to work :) wow that's long post, thanks for your patience if you got that far and any answers :) oh yeh and I use php/apache on shared hosting

    Read the article

  • <msbuild/> task fails while <devenv/> succeeds for MFC application in CruiseControl.NET?

    - by ee
    The Overview I am working on a Continuous Integration build of a MFC appliction via CruiseControl.net and VS2010. When building my .sln, a "Visual Studio" CCNet task (<devenv/>) works, but a simple MSBuild wrapper script (see below) run via the CCNet <msbuild/> task fails with errors like: error RC1015: cannot open include file 'winres.h'.. error C1083: Cannot open include file: 'afxwin.h': No such file or directory error C1083: Cannot open include file: 'afx.h': No such file or directory The Question How can I adjust the build environment of my msbuild wrapper so that the application builds correctly? (Pretty clearly the MFC paths aren't right for the msbuild environment, but how do i fix it for MSBuild+VS2010+MFC+CCNet?) Background Details We have successfully upgraded an MFC application (.exe with some MFC extension .dlls) to Visual Studio 2010 and can compile the application without issue on developer machines. Now I am working on compiling the application on the CI server environment I did a full installation of VS2010 (Professional) on the build server. In this way, I knew everything I needed would be on the machine (one way or another) and that this would be consistent with developer machines. VS2010 is correctly installed on the CI server, and the devenv task works as expected I now have a wrapper MSBuild script that does some extended version processing and then builds the .sln for the application via an MSBuild task. This wrapper script is run via CCNet's MSBuild task and fails with the above mentioned errors The Simple MSBuild Wrapper <?xml version="1.0" encoding="utf-8"?> <Project ToolsVersion="4.0" DefaultTargets="Build" xmlns="http://schemas.microsoft.com/developer/msbuild/2003"> <Target Name="Build"> <!-- Doing some versioning stuff here--> <MSBuild Projects="target.sln" Properties="Configuration=ReleaseUnicode;Platform=Any CPU;..." /> </Target> </Project> My Assumptions This seems to be a missing/wrong configuration of include paths to standard header resources of the MFC persuasion I should be able to coerce the MSBuild environment to consider the relevant resource files from my VS2010 install and have this approach work. Given the vs2010 msbuild support for visual c++ projects (.vcxproj), shouldn't msbuilding a solution be pretty close to compiling via visual studio? But how do I do that? Am I setting Environment variables? Registry settings? I can see how one can inject additional directories in some cases, but this seems to need a more systemic configuration at the compiler defaults level. Update 1 This appears to only ever happen in two cases: resource compilation (rc.exe), and precompiled header (stdafx.h) compilation, and only for certain projects? I was thinking it was across the board, but indeed it appears only to be in these cases. I guess I will keep digging and hope someone has some insight they would be willing to share...

    Read the article

< Previous Page | 120 121 122 123 124 125 126 127 128 129 130 131  | Next Page >