Search Results

Search found 21955 results on 879 pages for 'self host'.

Page 471/879 | < Previous Page | 467 468 469 470 471 472 473 474 475 476 477 478  | Next Page >

  • What does a media server need?

    - by DeaDSouL
    I want to buy a new PC to be used as a media server that streams HD movies, music to any wireless device on my network (laptop, desktop, TV, etc.). I was wondering about the PC hardware's specifications that I need to have so that it can work just fine? (It will be wireless) Is there any tutorial showing how it could be done? And the security - how can I secure it? As well as I need it to to host my emails. I mean I have many emails. And sometimes I'm using desktop, sometimes laptop.. so is it possible to make the server store a copy of my emails, attachments, all sent items (whether they were from desktop or laptop)?

    Read the article

  • Wordpress : Automatically transfer media files to Amazon S3

    - by Ron Ranieri
    I've been using VPS to host 7 Wordpress websites, most of them require big storage but very little RAM and traffic. So I'm thinking of moving the static files(uploads folder) content to Amazon S3 and I'm looking for the most viable solution to this. I want every website to have their own bucket and newly uploaded media files automatically uploaded to Amazon S3 without using plugin. I'm ok with cron job, for example the files were uploaded first to my server, then transferred to S3 and deleted from my server every 24 hour. Or is there any way for me to change the default upload directory to my S3 bucket without sacrificing any Wordpress functionality(resize/title etc)? What do you think the most efficient way to do this? Currently I'm looking at this plus cron job but I would like to know better option if it exist.

    Read the article

  • vmbuilder fails on chroot

    - by Bruce
    I am trying to install virtual machine with this command, but have no success: vmbuilder kvm ubuntu --verbose --suite precise --flavour virtual \ --part partitions.txt --ip 192.168.1.3 --hostname edb1 --arch amd64 \ -o --libvirt qemu:///system --user someuser --pass somepass \ --raw /home/virtual-machines/edb1.disk1.img \ --raw /home/virtual-machines/edb1.disk2.img \ --domain somedomain.com --mem 4096 --cpus 4 This is the error: ... I: Extracting xz-utils... I: Extracting zlib1g... W: Failure trying to run: chroot /tmp/tmp_JdKzu mount -t proc proc /proc , stderr: The host kernel is not original but modified by server provider. Why is the chroot needed for installation?

    Read the article

  • Using an iPad or Nexus 7 as an Ubuntu Second Monitor

    - by never4getthis
    Is there any way to use an iPad mini or a Nexus 7 as a second monitor through a cable? I dont want something like vnc, I need it to work through the usb cable (or any other cable). I imagine that the iPad will get a lot of hate around here, the reason why I am considering it is because of the larger screen (compared to the nexus 7). I think its self explanatory that I need this to work with ubuntu (as I am posting in this formum). Thanks for reading this highly unorganized, typo-flled (<--see what I did there?), un-cohesive post. Any help will be appreciated!

    Read the article

  • Is ignoring clients data during an automatic upgrading acceptable?

    - by A competent translator
    I recently faced a deletion of my calendar and notes data inside a third party application ( Horde ) on my online space. When I told the web hosting provider they replied that they were doing an upgrade to system software (!!) and client third party applications are not within their reach or backup policy. They don't even have a backup of my data. Is this practice acceptable from a web hosting provider, even for an individual clients ? I know that backing up my data is my responsibility but I anticipated that backup copies done by the host would be available when needed.

    Read the article

  • How to access shared folders in Ubuntu VM(Oracle Virtualbox) and link them to the home folder(Answered)

    - by Njihia
    I have configured a shared folder between the Windows host and Ubuntu guest. The folder mounts at start up but its empty(It also has a padlock sign.). I have to run the command below to access its content(the padlock sign disappears). sudo mount.vboxsf media ~/media How can i configure it to be run automatically at start up. I've tried adding to the start up programs but nothing happens. Am new to Linux so try to put your answer in a layman's language. Thanks.

    Read the article

  • How can I get started installing Buddycloud software to a hosted site?

    - by rujic
    I'm confused about how one hosts a Buddycloud (federated social networking software) channel. I downloaded the source code, but I'm not sure how to install it on my website. Most of the information seems to assume that the host is running their own server, but I have plain ol' paid web hosting. I know that you also need to set up an XMPP server to run Buddycloud, and I'm not sure how to do that on my hosted website—or if it's even possible. If it matters, my site is hosted on a Linux server through IX Web Hosting. I'd greatly appreciate it if you guys could point me in the right direction.

    Read the article

  • Is the structure used for these web pages a design pattern?

    - by aspdotnetuser
    I want to know if the structure for an ASP.NET website I'm working on uses a design pattern for it's web pages. If it is a design pattern, what is it called? The web pages have the following structure: UserDetails page (UserDetails.aspx) - includes UserDetailsController.ascx user control. UserDetailsController.ascx includes sub user controls like UserAccountDetails.ascx and UserLoginDetails.ascx etc Each sub user control contains a small amount of code/logic, the 'controller' user controls that host these sub user controls (i.e UserDetailsController.ascx) appear to call the business rules code and pass the data to the sub user controls. Is this a design pattern? What is it called?

    Read the article

  • Flickering/Flashing screen problem related to keyboard

    - by LeandroCR
    This will seems quite bizarre but that's a huge problem for me, I'm facing it for a long time since Ubuntu 12.04. My Ubuntu/Linux Mint/Debian (Unity/Gnome) always starts flickering after using the computer for a while, I always thought this had something to do with the video board. I have an Vostro 3550 (Radeon 6xxx/Intel Onboard). First I realized that when I was navigating trough an website, when i selected with the mouse a combo, the combo closed it self automatically then the screen starts flickering. The i realized when i was trying to make the login in the terminal (crtl + alt + f1) a strange character appears, like: ^@ repeatedly. When i keep CRTL pressed the system never flickers! I have absolutely no problem using Windows. What can i do? There is no such thing as Keyboard driver on Dell Support. I would love to use Ubuntu only, but this is very annoying. Problem happens on Ubuntu 12.04, .10 and 13.04 Linux Mint 15 (Cinnamon) Debian Live with Gnome

    Read the article

  • Why is my LAN not working properly ?

    - by ps06756
    I am using Ubuntu 12.04 as my OS. Till yesterday, my LAN was working fine , but after that my LAN is not working properly. What I mean is that it shows that the LAN wire is connected properly, but when I try to open a website it gets stuck on resolving host. I am using the web browser as Google-chrome. Now, when I manually switch off the wired connection and turn it on again, the LAN works perfectly till I open one or two webpages. After that the same problem repeats again. Can anybody, help me with a solution to this problem.

    Read the article

  • Static IP breaks Network connection

    - by Pasquale Tedesco
    I am somewhat new to Linux. I am setting up a Web server and installed Ubuntu server 14.04 on an vSphere ESXi 5.5 host. I am connecting perfectly fine to the internet when using DHCP but when I apply my Static settings I loose the ability to connect to websites. Whats weird is I can Ping google.com but if I try wget gooogl.com it hangs at connecting to google.com (google.com):173.194.43.32:80 and I get "failed: connection refused" But if I am set to DHCP the connection resolves perfectly. Thanks # This file describes the network interfaces available on your system # and how to activate them. For more information, see interfaces(5). # The loopback network interface auto lo iface lo inet loopback # The primary network interface auto eth0 #iface eth0 inet dhcp iface eth0 inet static address 10.1.2.25 gateway 10.1.2.2 netmask 255.255.255.0 search tsig.com dns-nameservers 10.1.2.13 10.1.2.30

    Read the article

  • How to purge old links in google from an old domain.

    - by jbcurtin
    Hey all, Recently, I uploaded a new site to an existing domain and I'd like to figure out how I can forward all links to said domain to a new domain. I'm looking for a wordpress solution if possible, but in the end I I seem myself writing a small header script that I will paste into ever directory's index file saying header('Location:http://xxx.yyy.zzz') Is there a cleaner way to do this without having to resort to managing the whole file structure? No, I do not have access to the apache runtime. Unfortunately it is a shared-host server. Thanks in advance.

    Read the article

  • Hosting a magnet link site which could possibly infringe copyrighted material?

    - by Griff
    I have for the last 3 months built a crawler, indexer and alot of other things for what started out to be a home project for indexing magnet links on the internet. As my project grew I have thought about releasing my collected data (which at the minute is on a public domain but with no access) to the public. Whatever the crawler sucks in goes in, and whatever the indexer decides to index gets indexed as it is a fully automated process. My question is as follows; Considering that most of the data that is collected from what I have built points to illegal copyrighted material (as most magnet links do) where would it be best to host such a site. I notice all of the already public torrent sites are hosted in India is this because there laws are less strict on copyright infringement? Have any of you hosted such a site, and if so what problems have you ran into? And as always any advice on being a webmaster for this type website?

    Read the article

  • How do I migrate web files from a Plesk 8 installation (on a slaved HDD) to a Plesk 10.4.4 installation?

    - by Ranger
    Due to Plesk 8 being at End of Support our host setup a new installation of RHEL and Plesk 10 on a new hard drive. They then slaved the old drive to the new so that we could migrate all our files using SSH. I am having challenges correctly migrating the sub domain files. The path to subdomain root folder in Plesk 10.4.4 is confusing as I don't know where to copy the files to. The path to the files on the slaved drive is: "/mnt/old-drive/var/www/vhosts/domain_name.com/subdomains/SUBDOMAIN_NAME/", meanwhile on the new installation I have "/var/www/vhosts/SUBDOMAIN_NAME.domain_name.com". There is an httpdocs folder in the '/var/www/vhosts/domain_name' folder but none in the '/var/www/vhosts/SUBDOMAIN_NAME.domain_name.com' folder. Where do I copy my subdomain files to? Please help.

    Read the article

  • Forum that integrates into CMS and has curated category pages with tagged threads

    - by user6172
    I'm looking for a forum that meets these requirements: Login using Facebook/Twitter/OpenID etc. User profiles with reward system Voting/thumbs up function Categories and tags for sorting threads Custom category pages with moderated static header Embeddable threads and categories (For example, a whole category or single thread can be integrated into wordpress) API to users, discussions etc. I've looked at forums like Vanilla, Disqus, OSQA etc, but none seem to match the above "hybrid criteria". Hosted or self-hosted doesn't matter but I'm really looking for something that can be integrated into an existing CMS to replace comments while at the same time have curated category pages and user profiles. Thanks.

    Read the article

  • How to create a temporary staging server on my home machine [closed]

    - by Homunculus Reticulli
    Possible Duplicate: What things required to host a website at home I want to create a temporary staging server which can be accessed (i.e. via browser) by other people that I want to show the website to (a business partner who is half way accross the world). IIRC, my ISP issues dynamic addresses so I may need to register with a (DNS server?) - not sure about this. Although I'm a software developer, I don't know much about the hardware side of things - and would appreciate help in getting me setup so I can show a website to a business partner. Here are the relevant details: Web server: Apache 2.2 OS: Ubuntu 10.0.4 LTS modem/router: ZyXel P-600

    Read the article

  • SSD becomes hot, disk failure warning

    - by Aegluin
    I have a two weeks old SSD (Kingston SSDnow 64GB). Yesterday, the computer shutdown twice and after rebooting I was bombarded with disk failure warnings. I usually take such warnings serious (and backed up), but skeptical. After cooling down, the laptop boots again and the only red Smart value was the temperature (Ubuntu did not show the temperature of failure, but the at that time 29°). After refreshing the Smart status and doing a "self test", everything is green. Before contacting Kingston support, I would like to know whether it could be due to a software issue: Is it possible that it is false alarm, and how can I check? I installed Ubuntu 12.04 32bit and took care of alignment. I supposed Ubuntu set up with optimal settings for SSDs, how can I check that there was no mistake? The current temperature is around 40-56°. Is such a temperature abnormal for SSDs? Output of sudo smartctl --all /dev/sda: http://pastebin.ubuntu.com/1175940/

    Read the article

  • Could JQuery and similar tools be built into the browser install?

    - by Ozz
    After reading another question about JQuery and CDN's, is it feasible for tools like JQuery to "come with" the browser, thus reducing/eliminating the need for the first download from a CDN, or from your own host server. JQuery files specifically are pretty small, so you could easily have a number (all?) of the different versions as part of a browser install. Now fair enough, this would increase install footprint, download time for the browser itself. Then sites could check "local" first, before CDN (which then caches), before finally defaulting to downloading from the website server itself. If this is feasible, has it been done, and if not, why hasn't it be done?

    Read the article

  • Advice on career change [on hold]

    - by user72406
    I am a software developer(sharepoint developer) with total 3 years of experience.Though using out of the box sharepoint excites me, i have aversion to code. i kinda cant put my brains in coding. i have my eyes on MBA now(only reason is to get rid of code). I wanted to rethink on this and need your expert advice and suggestions. As i am working currently in a hot skill(sharepoint 2013) technology, is it a good thought to go ahead and do MBA and change my career, or should i start liking what i do?? THough it is supposed to be a self decision, i needed views of experts to help me carve my niche. Regards Kiran

    Read the article

  • How would I know if my OS is compromised?

    - by itsols
    I had opened a php folder from a friend's web host. I run it on mine to fix some bugs. Then I tried attaching the code to be emailed and GMAIL stated that the attachment was infected by a virus. Now I'm afraid if my Apache or OS (12.04) is infected. I checked the php files and found a base64 encoded set of code being 'eval'd at the top of each and every php file. Just reversing it (echo with htmlspecialchars) showed some clue that there were sockets in use and something to do with permissions. And also there were two websites referred having .ru extensions. Now I'm afraid if my Ubuntu system is affected or compromised. Any advice please! Here's my second run of rkhunter with the options: sudo rkhunter --check --rwo Warning: The command '/usr/bin/unhide.rb' has been replaced by a script: /usr/bin/unhide.rb: Ruby script, ASCII text Warning: Hidden directory found: /dev/.udev Warning: Hidden file found: /dev/.initramfs: symbolic link to `/run/initramfs'

    Read the article

  • $ON_USER returning root instead of $USER

    - by Nathanel Titane
    Hello everybody! With Natty coming out soon, I've been at work updating my deployment and self-config script to make my desktop on 11.04 run and look the way I want it to. One bummer is that dbus seems to have changed and does not permit, in the same manner Lucid and Maverick did, the authentication of the current user by terminal call using grep and cat. Ideally, to run the script, I would sudo -s and then launch it as # chmod +x install && ./install Instead of returning my user name.. it now returns root and applies changes to the root profile and aborts whenever paths do not correspond. Here is my script header: #!/bin/bash ON_USER=$(echo ~ | awk -F'/' '{ print $1 $2 $3 }' | sed 's/home//g') export $(grep -v "^#" ~/.dbus/session-bus/`cat /var/lib/dbus/machine-id`-0) if sudo -u $ON_USER test -z "$DBUS_SESSION_BUS_ADDRESS" ; then eval `sudo -u $ON_USER dbus-launch --sh-syntax --exit-with-session` fi RELEASE=$(lsb_release -cs) How could I make it return the actual user now that natty is coming? Thanks for the help

    Read the article

  • How should I choose quadtree depth?

    - by Evpok
    I'm using a quadtree to prune collision detection pairs in a 2d world. How should I choose to what depth said quadtree is calculated? The world is made mostly of moving objects1, so the cost of dispatching the objects between the quadtree cells matters. What is the relationship between the gain from less collision checking and the loss from more dispatching? How can I strike a balance that performs optimally? 1 To be completely explicit, they are autonomous self-replicating cells competing for food sources. This is an attempt to show my pupils predator-prey dynamics and genetic evolution at work.

    Read the article

  • FatCow and iPage real deal?

    - by Tribbey
    I need a cheap host that I can upgrade if needed for a startup. From searching, FatCow and iPage seem reliable and inexpensive web hosting services with a Unix OS and good bandwidth + disk space. I suspect they were bought from the same company. They propose to offer unlimited bandwidth/disc space and perks like AdSense/Facebook credit however, they're packages range from 1-3 years @ 3.15 USD/mo, there isn't a monthly package but they do allow you to cancel at anytime and there servers run on wind-mill generated energy which is a plus. I was suspicious because I couldnt seem to find one negative review about them, just affiliate pages until I read a review explaining they're strict policies on copywritten data. Has anyone experience with one of these two hosts?

    Read the article

  • How to direct a Network Solutions domain name to an html website hosted on Google Drive? [on hold]

    - by Air Conditioner
    To begin with, I'd wanted to take advantage of HTML, CSS, and so on to build a website that looks and works just as I'd like it to. I took a look around on how I could make that work, and I soon saw a lifehacker article showing that its possible to host website files on google drive. I then made sure that the folder containing the files was shared publicly throughout the web, and I now have a working 'google drive hosted' domain for the website. However, I did want to have the custom domain, and so I registered one with network solutions. So now, I'm curious on how I should direct my Network Solutions domain to the index.html I'm hosting on google drive. Would anyone have an Idea?

    Read the article

  • How do I configure multiple domain names on my IIS server? [closed]

    - by Dillie-O
    We have a few websites that we are running on one instance of IIS that need to be mapped for each of their domain names. For example. Site A has the domain name coolness.com Site B has the domain name 6to8Weeks.com Site C has the domain name PhatTech.com When I look at the "Web Site Identification" section of the IIS configuration window, I notice that I can specify an IP address and port, but if I click the Advanced button, I can also configure the site based on host header values as well. How do I configure each site in IIS? Ideally I would like them to all be able to listen to port 80, so I don't have weird URLs, but I'm not sure if I do this using headers, IP addresses, both, or something else.

    Read the article

< Previous Page | 467 468 469 470 471 472 473 474 475 476 477 478  | Next Page >