Search Results

Search found 32520 results on 1301 pages for 'local machine'.

Page 618/1301 | < Previous Page | 614 615 616 617 618 619 620 621 622 623 624 625  | Next Page >

  • Pointing domain away from google

    - by redbeard
    I've already asked this question on google apps support, didnt fet an answer yet, and I am in a bit of a hurry. Need to have that website up by monday. Question: We have a domain registered at 101domain, we used a domain just for google apps, for email, and some services, and domain was never used for website. Now we pointed domain to the local hosting service, set everything like it should be, and pointed back to google from hosting service for gmail and gmail works perfectly. Problem is that we tried to set up a website at hosting service but domain still points to google apps login. Everything looks like it should, we recently bought 2 more domains that are set up the same (we use google apps on them too), the only difference tha I can think of is that we didn't use for apps first. Could this be because CNAME lags behind DNS server change?

    Read the article

  • How to transform mesh components?

    - by Lea Hayes
    I am attempting to transform the components of a mesh directly using a 4x4 matrix. This is working for the vertex positions, but it is not working for the normals (and probably not the tangents either). Here is what I have: // Transform vertex positions - Works like a charm! vertices = mesh.vertices; for (int i = 0; i < vertices.Length; ++i) vertices[i] = transform.MultiplyPoint(vertices[i]); // Does not work, lighting is messed up on mesh normals = mesh.normals; for (int i = 0; i < normals.Length; ++i) normals[i] = transform.MultiplyVector(normals[i]); Note: The input matrix converts from local to world space and is needed to combine multiple meshes together.

    Read the article

  • Getting to math applications gradually

    - by den-javamaniac
    I'm currently getting a formal degree related to computation, in particular my current focus is numerical programming, scientific computing and machine learning. I'd love to apply that knowledge in game dev and expand it with statistics, probability theory, and graph theory (probably even linear algebra). The question is: which spheres of gamedev are filled with such math stuff, is it possible to advance in those without being a part of a group of people and how to get to it gradually? P.S.: I've got experience with commercial java dev and am getting my hands on C/C++ at the moment, however, I'm opened to go ahead and try Unity3D and etc.

    Read the article

  • Wordpress asks for FTP credentials, even though owner of the whole /var/www is www-data

    - by dmt0
    I'm trying to get Wordpress to run on my local Ubuntu 12.10 installation. When I try to install themes/plugins, it asks me for FTP credentials. I've been trying to get this to work for 2 days now. Everywhere on the web it says you should change the owner:group of your Wordpress directory to whateverowner:group runs your apache server - in my case www-data:www-data. I've done this, and even tryed setting permissions to the whole /var/www directory to 777, restarting apache and the whole system. But wordpress is still asking me for the FTP credentials. What else could this be?

    Read the article

  • Using a back-end mechanism to copy files to DB and notify the application

    - by BDotA
    This Scenario: User copies large files to a local folder. I want to watch that folder and when a new file is dropped then go and copy it to Database, so later when coping is done I can actually use it in my application. ( A C# WinForms App). It would be awesome to also find a way to somehow get notified in the Application that hey copying the file to DB is finished and ready for use... I am using C#.net, Windows... What solutions/architecture do you suggest for this? For example having a windows service running all the time watching that folder, when something copied goes and write it to DB ... then how about getting notified? MSMQ is something I can use? don't know much about it yet. Thanks.

    Read the article

  • RTL8192SU-based Wi-Fi adapter disconnects permanently

    - by leventov
    I've already tried all possible (http://ubuntuforums.org/showpost.php?p=10129571&postcount=43) solutions, no progress. I'm in despair. In Windows (on the same machine) this adapter works stably. Device: Trendnet TEW-649UB. System details: Ubuntu 11.10; leventov@leventov-ubuntu:~$ uname -a Linux leventov-ubuntu 3.0.0-12-generic #20-Ubuntu SMP Fri Oct 7 14:56:25 UTC 2011 x86_64 x86_64 x86_64 GNU/Linux leventov@leventov-ubuntu:~$ lsusb ... Bus 002 Device 002: ID 0bda:8172 Realtek Semiconductor Corp. RTL8191S WLAN Adapter ... leventov@leventov-ubuntu:~$ dmesg | grep 8712 #current driver [ 8.146510] r8712u: module is from the staging directory, the quality is unknown, you have been warned. [ 8.147113] r8712u: DriverVersion: v7_0.20100831 [ 8.147124] r8712u: register rtl8712_netdev_ops to netdev_ops [ 8.147127] r8712u: USB_SPEED_HIGH with 4 endpoints [ 8.147478] r8712u: Boot from EFUSE: Autoload OK [ 8.551272] r8712u: CustomerID = 0x0000 [ 8.551275] r8712u: MAC Address from efuse = 00:14:d1:6c:52:19 [ 8.551625] usbcore: registered new interface driver r8712u [ 9.501351] r8712u: Loading firmware from "rtlwifi/rtl8712u.bin" [ 10.160471] r8712u: 1 RCR=0x153f00e [ 10.161241] r8712u: 2 RCR=0x553f00e leventov@leventov-ubuntu:~$ lsmod | grep 8712 r8712u 189049 0 t

    Read the article

  • Which hidden files and directories do I need?

    - by Sammy Black
    In a previous question, I explained my situation/plan: backing up home directory on external drive, reformatting laptop drive, installing 14.04, putting home directory back. (It hasn't happened yet because I can't seem to find the down time, in case things aren't working right away.) It occurred to me that maybe I don't want all of those hidden files and directories (e.g. .local/share/ubuntuone/syncdaemon/, .cache/google-chrome/, etc.) Just judging by the amount of time in copying, I can tell that some of these hidden directories are large. Question: Are there any hidden directories that I obviously don't need/want when I have the laptop running an updated distribution? Will they cause conflicts? (I plan on copying the backed-up directory tree back onto the laptop with the --no-clobber option.)

    Read the article

  • A Patent for Workload Management Based on Service Level Objectives

    - by jsavit
    I'm very pleased to announce that after a tiny :-) wait of about 5 years, my patent application for a workload manager was finally approved. Background Many operating systems have a resource manager which lets you control machine resources. For example, Solaris provides controls for CPU with several options: shares for proportional CPU allocation. If you have twice as many shares as me, and we are competing for CPU, you'll get about twice as many CPU cycles), dedicated CPU allocation in which a number of CPUs are exclusively dedicated to an application's use. You can say that a zone or project "owns" 8 CPUs on a 32 CPU machine, for example. And, capped CPU in which you specify the upper bound, or cap, of how much CPU an application gets. For example, you can throttle an application to 0.125 of a CPU. (This isn't meant to be an exhaustive list of Solaris RM controls.) Workload management Useful as that is (and tragic that some other operating systems have little resource management and isolation, and frighten people into running only 1 app per OS instance - and wastefully size every server for the peak workload it might experience) that's not really workload management. With resource management one controls the resources, and hope that's enough to meet application service objectives. In fact, we hold resource distribution constant, see if that was good enough, and adjust resource distribution if that didn't meet service level objectives. Here's an example of what happens today: Let's try 30% dedicated CPU. Not enough? Let's try 80% Oh, that's too much, and we're achieving much better response time than the objective, but other workloads are starving. Let's back that off and try again. It's not the process I object to - it's that we to often do this manually. Worse, we sometimes identify and adjust the wrong resource and fiddle with that to no useful result. Back in my days as a customer managing large systems, one of my users would call me up to beg for a "CPU boost": Me: "it won't make any difference - there's plenty of spare CPU to be had, and your application is completely I/O bound." User: "Please do it anyway." Me: "oh, all right, but it won't do you any good." (I did, because he was a friend, but it didn't help.) Prior art There are some operating environments that take a stab about workload management (rather than resource management) but I find them lacking. I know of one that uses synthetic "service units" composed of the sum of CPU, I/O and memory allocations multiplied by weighting factors. A workload is set to make a target rate of service units consumed per second. But this seems to be missing a key point: what is the relationship between artificial 'service units' and actually meeting a throughput or response time objective? What if I get plenty of one of the components (so am getting enough service units), but not enough of the resource whose needed to remove the bottleneck? Actual workload management That's not really the answer either. What is needed is to specify a workload's service levels in terms of externally visible metrics that are meaningful to a business, such as response times or transactions per second, and have the workload manager figure out which resources are not being adequately provided, and then adjust it as needed. If an application is not meeting its service level objectives and the reason is that it's not getting enough CPU cycles, adjust its CPU resource accordingly. If the reason is that the application isn't getting enough RAM to keep its working set in memory, then adjust its RAM assignment appropriately so it stops swapping. Simple idea, but that's a task we keep dumping on system administrators. In other words - don't hold the number of CPU shares constant and watch the achievement of service level vary. Instead, hold the service level constant, and dynamically adjust the number of CPU shares (or amount of other resources like RAM or I/O bandwidth) in order to meet the objective. Instrumenting non-instrumented applications There's one little problem here: how do I measure application performance in a way relating to a service level. I don't want to do it based on internal resources like number of CPU seconds it received per minute - We need to make resource decisions based on externally visible and meaningful measures of performance, not synthetic items or internal resource counters. If I have a way of marking the beginning and end of a transaction, I can then measure whether or not the application is meeting an objective based on it. If I can observe the delay factors for an application, I can see which resource shortages are slowing an application enough to keep it from meeting its objectives. I can then adjust resource allocations to relieve those shortages. Fortunately, Solaris provides facilities for both marking application progress and determining what factors cause application latency. The Solaris DTrace facility let's me introspect on application behavior: in particular I can see events like "receive a web hit" and "respond to that web hit" so I can get transaction rate and response time. DTrace (and tools like prstat) let me see where latency is being added to an application, so I know which resource to adjust. Summary After a delay of a mere few years, I am the proud creator of a patent (advice to anyone interested in going through the process: don't hold your breath!). The fundamental idea is fairly simple: instead of holding resource constant and suffering variable levels of success meeting service level objectives, properly characterise the service level objective in meaningful terms, instrument the application to see if it's meeting the objective, and then have a workload manager change resource allocations to remove delays preventing service level attainment. I've done it by hand for a long time - I think that's what a computer should do for me.

    Read the article

  • SQL Auto Close Options

    - by Dave Noderer
    Found an interesting thing that others have run across but it is the first time I’ve seen it. A customer emailed to say that the SQL 2008 db that I had helped him with seemed to be going into recovery mode on a regular basis while watching the SQL Management Studio screen. Needless to say he was a bit nervous and about to take some drastic steps. Eventually he found that the Auto Close option was set to true. When this is set to true, the database automatically closes all connections and unlocks the mdf file after 300 milliseconds. When a new connection is made it spins backup… Great for xcopy deployment on a client machine but not a multi-user server based application. So the warning… if you have started a database with SQL express and then move it to a production SQL server, make sure you check that the Auto Close option is set to false. See options screen below:

    Read the article

  • Torrent clients suddenly stopped downloading

    - by Vasilis Baltikas
    A few days ago I noticed, that Transmission in my Ubuntu 10.04 machine suddenly couldn't download anything anymore. To overcome this I have uninstalled and reinstalled Transmission and tried downloading with other clients (Deluge, Vuze) well seeded torrents (ubuntu iso images for example) without success. On the same computer I have also installed Ubuntu 10.10 and Windows 7, which I rarely use. What makes the problem I encounter weirder, is the fact that downloading via torrents works fine in my Ubuntu 10.10 and Windows partitions but not in Lucid Lynx. Browsing the web for similar problems didn't give me answers. Any help would greatly appreciated.

    Read the article

  • Trip Report: Scottsdale Arizona Event Conference Planning

    Monday May 10th, 2010 represents my largest and best event to-date, that I have ever hosted. For the last seven years, I’ve grown a technical audience around Microsoft programming / developers in the Phoenix area. Some travel from California, New Mexico, and other parts of Arizona, but for the most part the crowd is local. In years past, this all day event has been as small as 300 folks, and as large as 500, and I’ve tried been to venues around the valley some better than others. This...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Skype no sound on Kubuntu 13.10

    - by Michael Aquilina
    I just performed a fresh install of Kubuntu 13.10 on my machine. Everything is working great except for Skype. I cannot get any form of audio playback in Skype. In the sound settings panel I get a bunch of different sound sources, none of which work! At the moment I have set it to "sysdefault (unknown)" I installed it using the deb package found on the official website. My phonon backend is using phonon-gstreamer. When running skype from the terminal I get the following error messages: ALSA lib control.c:953:(snd_ctl_open_noupdate) Invalid CTL plughw:CARD=PCH ALSA lib pcm_dmix.c:1022:(snd_pcm_dmix_open) unable to open slave ALSA lib pcm_dmix.c:1022:(snd_pcm_dmix_open) unable to open slave ALSA lib pcm_dmix.c:1022:(snd_pcm_dmix_open) unable to open slave ALSA lib pcm_dmix.c:1022:(snd_pcm_dmix_open) unable to open slave ALSA lib pcm_dmix.c:1022:(snd_pcm_dmix_open) unable to open slave ALSA lib pcm_dmix.c:1022:(snd_pcm_dmix_open) unable to open slave Is this a known problem or has anyone experienced the problem and managed to solve it?

    Read the article

  • Windows 7 Virtual PC - &ldquo;RPC server unavailable&rdquo;

    - by Kelly Jones
    I use Windows 7 Virtual PC on my current project and I often bring home the files, so I can work some in the evenings.  Since my VHDs are large, I’ll only copy the undo disks, saved state, and virtual machine config files from my external drive.  I copy them to a small portable drive and once I get home, I’ll copy them to a large external drive. I’ve done this for over a year, but recently I started getting an error when I tried to start the VPC after the copying was finished.  It would open the initial window with the progress bar, but eventually the bar would stop, turn red, and then the error “RPC server unavailable” would appear.  When I first started seeing these, I’d try again, but no luck. After some testing, it turns out that my small portable drive is apparently going bad, so it was corrupting the files.  Lucky for me, that I never overwrote my good copies with corrupted copies, at least not at both the office and at home.

    Read the article

  • Where to draw the line between front end and back end

    - by Twincascos
    I was recently contracted to develop a smarty theme for an automated SOHO phone answering service. The team who had built the backend wouldn't allow me access to any of the back end nor tell me anything about it, their smarty set up, smarty plugins, data base interface api, server set-up, nothing. Nor could I have access to the server nor a beta domain, basically zero co-operation. So I set up a local server with Smarty and built the template based on what I guessed would be their best practice, commented my code like crazy, wrote all the needed javascript, css, and template files. Then I sent them packaged to the backend team and hoped for the best. With half of a project team failing to cooperate or even communicate I am now concerned that they may reply saying that everything is wrong and they may refuse to implement the new front end. I'm curious to know if others encounter this type of situation and what you may have done to protect yourselves.

    Read the article

  • Is there a way to use a SSH connection to access SMB or UPnP files without setting up a VPN?

    - by Michael Chapman
    What I'm trying to do is set up a SSH key that only gives access to certain directories, for security reasons I don't want it to have full access to my SSH server. I already have the ability to access the directories I need over my local network (right now using SMB, although I also used UPnP for awhile). I need, however, to be able to access those files securely over the internet from both Ubuntu and Windows machines. I'm somewhat new to SSH and not sure what the best approach to solving my problem is. If anyone knows how I can do this or where I can find a detailed tutorial I'd be grateful. And as always if anything is confusing or if there are any comments or corrections please let me know.

    Read the article

  • Uninstall Calibri font, installed from Windows' font directory

    - by donkeydown
    I use e PC with both Ubuntu and Windows-7 OSs. I tried to install the Calibri font for Ubuntu in this way: I did open the calibri.ttf file in the Windows/Fonts directory using Font Viewer and clicked on the Install button. Now I need to disintall that font but I can't. Font Viewer shows me "Calibri Regular" font is installed but does not allow me to disintall it. Font Manager does not show Calibri in the font list. Character Map does not show Calibri in the font list. Ubuntu Software Center and Synaptic don't find anything like "calibri". There is no calibri file in those directories: usr/share/fonts usr/local/share/fonts ~/.fonts The font is visible to LibreOffice, Chrome, Firefox.

    Read the article

  • Arabic disappeared after 12.04 upgrade!

    - by Aboubakr
    Well, I was amongst the 12.04 Beta upgraders, and since then I've lost the ability to write in Arabic, and I've been using Ubuntu since 2008 as an only OS without any issues, and been upgrading since then as well, except on this machine, which received one upgrade from 11.10 to 12.04 and it got messed up. I've added Arabic like usual, but it doesn't change with the keyboard's short-cut, and when I do it manually with the Mouse, then it just doesn't work, and it keeps writing in English instead. I've tried to install some iBus things, and added Arabic-kbd (m17n) but it still remains messy, let alone not having the same layout, and all I want is to get to NORMAL. So, please, is there any way to reset or initialize these keyboard related settings, so I can get back to normal and stop using the Mac just to type in Arabic, or so often using XP over Vbox? And please, no Re-install option! I just can't backup all my work right now, and there are a lot of tasks waiting for me to get them done. Thanks for any kind of support :)

    Read the article

  • GRUB- error: no such partition grub rescue and Error: No default or UI configuration directive found boot > on pendrive

    - by Ash
    I have dell inspiron, previously I installed Ubuntu 11.10 on my Windows 7 and made it dual boot. But since I want to upgrade my Ubuntu version and change the partition spacing, I deleted 11.10 partition directly and extended my hardrive space (Windows + Ubuntu) at that moment everything was fine. Then I prepared a 12.04 32bit USB and installed it . It was installed but isn't showing dual boot option like 11.10 and my machine directly boot into Windows 7. So instantly i again deleted my 12.04 partition . Now I login into Windows 7 but whenever I put USB ( with 12.04 ) to boot from it, I am facing error of "no such partition grub rescue" even though I try to put lower version(11.04) it showing another error "Error: No default or UI configuration directive found boot " I have reinstall Windows 7 and reformat all partition, still I am facing same error :(

    Read the article

  • PXE-boot for Ubuntu 10.04 Desktop version

    - by omkar
    My aim is to Install the Ubuntu 10.04 desktop version on a remote machine using PXE-BOOT. I'm trying to apply the steps given in PXE-BOOT for Ubuntu 9.10 . (I know this is for 9.10). In "Step 8. Setting Up Boot Files", it says cp -a /media/cdrom/install/netboot/* /tftpboot/ but I wasn't able to find the netboot folder in my Ubuntu 10.04 Desktop CD-Rom. Does that folder only exist in Ubuntu Server edition? Is it possible to do this in the Desktop edition?

    Read the article

  • web services, J2EE, spring, DB integration project ideas- maybe data mining related?

    - by sj88
    Hey guys, I am a graduate CS student (Data mining and machine learning) and have a good exposure to core JAVA (3 years). I have read up a bunch of stuff on Design patterns J2EE Web services( soap and rest) spring and hibernate Java Concurrency - advanced features like Task and Executors. I would now like to do a project combining this stuff (over my free time of corse) to get a better understanding of these things and to kind of make an end to end software (to learn the best design principles etc + svn, maven). Any good project ideas would be really appreciated. I just wanna build this stuff to learn so I dont really mind re-inventing the wheel. Also, anything related to data mining would be an added bonus (fits with my research) but absolutly not necesary (since this project is more to learn to do large scale software developement)

    Read the article

  • Value of Internationalization in the iPhone App store?

    - by hotpaw2
    I have several iOS/iPhone apps that have been continually selling in small amounts in over 2 dozen different countries, even though the app UIs and all the store descriptions are only in English. In a few countries where English is not the official or native language, a few apps are selling far better than is proportionate for those country's population size compared with the U.S. So why Internationalize apps? What kind of increase, if any, in sales might a typical app see if it is Internationalized into given local languages? Which major languages might be likely to see the greatest improvement in app sales or downloads due to a localized app description?

    Read the article

  • Participez aux ateliers de certification Oracle à Paris le 30 octobre & 09 novembre 2012

    - by mseika
    Participez aux ateliers de certification Oracle à Paris le 30 octobre & 09 novembre 2012 Remportez la préférence de vos clients et prospects grâce à vos spécialisations Oracle ! Dans la continuité de votre démarche vers la certification Oracle, nous vous proposons 2 demi-journées "spéciale ateliers de certification" à Paris. Réservez votre matinée du 30 octobre ou du 09 novembre prochains pour passer les certifications indispensables à votre entreprise pour être spécialisée.Les ateliers auront lieu à Paris Saint Lazare de 9h à 12h30 au :Centre M2i20 rue d'Athènes75009 PARISNe manquez pas cette occasion, de nombreux ateliers au choix vous sont proposés. Attention, le nombre de places est limité. Programme des ateliers de certifications :- Oracle Software : Oracle Database 11g, Database Security, Data Integration, Data Warehousing, Oracle Business Intelligence Foundation, Exadata Database Machine, Exalogic Elastic Cloud, SOA... - Oracle Hardware : Oracle Linux, Oracle Solaris, SPARC Entry & Midrange, SPARC T-Series Servers, Stockage Unifié, Virtualisation Les ateliers seront suivis d'un déjeuner. Des pré-requis sont nécessaires pour passer ces examens en ligne.Vérifiez-les

    Read the article

  • SSL Certificate is Untrusted... sometimes

    - by dragonmantank
    Web Designer I'm working with signed up a new client that needed an SSL certificate. We went to namecheap.com and purchased on from Comodo. Got all the needed files and set it up in ISPConfig. To test we used Windows 7 running IE8, Firefox 3.6, and Chrome 12, and then on OSX with Firefox 4, Safari 5, and Chrome 13. All of them worked fine. The client is getting 'This connection is untrusted' in Firefox 4 and 5. Safari works fine on their machine. On my machines and the designer's machines all works with no errors. I had the client forward me the info for the certificate that Firefox has and the fingerprints match up. I have an old Windows 2000 VM with IE6 and Chrome and those work just fine as well. Any ideas on what else to check or do? The server is running Debian 5.0, up-to-date, with Apache 2 and ISPConfig 3.3

    Read the article

  • Ubuntu installation accuses false RAID

    - by rOim
    I'm trying to install a Ubuntu 12.04/Win7 dual boot on my machine (with 2 750GB HDs). Problem is, during the partitioning, the installation says I'm using RAID, and shows only one HD, with 1.5TB. I have, however, disabled raid on the setup, and installed Windows 7 only on HD 1, and want to keep Ubuntu to HD 2. I'm afraid to "resize" the partitions using the installer since that could mess my win7 installation. I did try to use the alternate installer, but had the same results.

    Read the article

  • In centralized version control, is it always good to update often?

    - by janos
    Assuming that: You are in a team developing some software. Your team is using centralized version control in the development process. You are working on a new feature which will surely take several days to complete, and you won't be able to commit before that because it would break the build. Your team members commit something every day that affects some of the files you're working with for your fancy new feature. Since this is centralized version control, you will have to update your local checkout at some point: at least once right before committing the new feature. If you update only once right before your commit, then there might be a lot of conflicts due to the many other changes by your teammates, which could be a world of pain to resolve all at once. Or, you could update often, and even if there are a few conflicts to resolve day by day, it should be easier to do, little by little. Can we say that it is always a good idea to update often?

    Read the article

< Previous Page | 614 615 616 617 618 619 620 621 622 623 624 625  | Next Page >