Search Results

Search found 25785 results on 1032 pages for 'oracles solution for soc'.

Page 553/1032 | < Previous Page | 549 550 551 552 553 554 555 556 557 558 559 560  | Next Page >

  • LDAP loginShell on platforms with different paths

    - by neoice
    I'm using LDAP to deal with users and authentication across my network. I'm now adding FreeBSD hosts and have hit a problem with login shells. on Linux, shells tend to be in /bin/$shellname, so setting my login shell in LDAP to /bin/zsh works perfectly. on FreeBSD, /bin/zsh doesnt exist, I need to use /usr/local/bin/zsh. is there a solution to this? I imagine I might be able to make some sort of login-shell.sh script that LDAP passes out as the "shell" and then use the script to determine the actual shell for the user, but I'm not a fan of that idea. I'm using Debian and FreeBSD, both with a standard OpenLDAP/PAM/nss setup. edit: it looks like using /bin/sh and adding an exec $shell to .profile would "work", but that doesnt scale very well.

    Read the article

  • HP Presario CQ 61-322ER (VV884EA) Wi-Fi hang up!

    - by qgrabber
    I have my new laptop and don't have Windows XP drivers for it. I found that it contains the Broadcom BCM4310 chip, but when I install any Broadcom driver my laptop hangup on installing bcm5*.sys driver. Only power-off button make any effect. After reboot the device list (Device Manager) contains Broadcom WLAN adapter, but it is marked as disabled, for some hardware errror! Also if I disable device before, and install driver - then - all is OK! But when I try to enable it, Windows hang up anyway (no speaker beep, no mouse input, no keyboard input - nothing) What is the solution?

    Read the article

  • Which hard disk drive is which?

    - by djeikyb
    I want to know which hard disk drive corresponds to which device path. It's trivial to match the hard disk stats (brand, size) with the dev path, but I want more. I want to know which drive is which inside my case. What's a good way to go about getting this info? Rules: I am a lazy bum. I don't want to tear apart my server to remove all the drives, then add back one by one. A reboot is acceptable. The drives are inconveniently scrunched together in the case. All label information is hidden. The case can be opened. Most disks are SATA, so theoretically hot swappable. Unplugging cables is fair game. Bonus: For a cli only solution. I'll award answer to the best/easiest gui or cli answer, and give a bounty to the next-best answer of the other kind. Or maybe the other way around, because the bounty is worth more points.

    Read the article

  • Syncing Large Directories/Filesystems using USB Drive

    - by Alan Lue
    Does anyone have a solution for syncing large directories/filesystems using just a USB flash drive (and specifically without using a network connection)? The objective is simply to sync a user directory between two computers. The contents of the user directory could amount to a large quantity of data—say, a quantity larger than could be stored on any single USB drive—but the aggregate size of changes that must be propagated by a single sync could easily fit on a USB drive. As an example, suppose a user directory is already synchronized between a desktop and a laptop computer. Here's a use case: Some changes are made in the user directory on the desktop. We mount a USB drive onto the desktop and copy whatever changes need to be applied to the laptop user directory in order to synchronize the desktop and laptop user directories. We now mount the USB drive onto the laptop and apply the changes. The desktop and laptop user directories are now synchronized. Any ideas? Alan

    Read the article

  • Arch Linux doesn't mount two USB drive?

    - by unixbhaskar
    One is USB modem which is connected and working and problematic one is a plain vfat USB stick which contain data ....that is the fellow not mounting :( I have tried to see it by fdisk ....it doesn't mount autometically... Is there any UDEV rule for that?? Because I have put a udev rule to remove the usb-storage thing( I had to.. otherwise the usb modem wont get connected... it waits for storage to relase the port) Any idea and solution would be greatly appreciated PS: I am running Arch distribution. Thanks

    Read the article

  • Automate opening HTML and printing to PDF

    - by craigpatik
    I need a way to automate the following process in Windows 7: Open an .html file in Internet Explorer Print to PDF Save the PDF with a patterned file name (i.e., original_name_YYYY-MM-DD.pdf) Ideally, I could drag and drop several files or open a whole folder of files at once and a PDF would be created for each one. A command line solution is also acceptable. The files have to be opened in the browser because parts of the page are rendered with JavaScript on page load. In other words, if you simply right-click on the file in Explorer and choose "print", the resulting file is not the same because the JS didn't run. If it helps, Internet Explorer can be set as the default browser, and a PDF printer can be set as the default printer.

    Read the article

  • Certain websites won't load with DHCP server

    - by Simon E
    Hi, I recently configured a server using Windows Server 2008 to take over from my router using DHCP Server. I disabled my router from assigning IPs too. All works well, except for certain sites that refuse to load e.g. "Firefox could not find www.amazon.co.uk" http://www.google.co.uk/ http://www.amazon.co.uk/ http://www.jobsite.co.uk/ I think it may have something to do with co.uk domain. If I reboot the server everything goes back to normal but of course thats not a good solution to do that every other day. Does anyone know of this problem or what it might be - I am very new to server technology.

    Read the article

  • Sharepoint 2007: Edit vs Read Only Mode

    - by user29116
    Sorry about the title, dont' really know what it should be. If I open a doc in read only mode I'm able to press save and then it opens up a save as box and the default directory is the directory on the sharepoint server and if you press save you save it to the server. This actually makes the whole process not really "read only" mode since I could actually update the document. Is there a way to prevent this from happening so that if someone chooses read only there is no way possible to updload any changes back to the sharepoint site? Also, it has been suggested as a solution to get rid of the edit/read only option so that people have to check out the document. Is there a way to remove the edit/read only option on documents?

    Read the article

  • How do 2D physics engines solve the problem of resolving collisions along tiled walls/floors in non-grid-based worlds?

    - by ssb
    I've been working on implementing my SAT algorithm which has been coming along well, but I've found that I'm at a wall when it comes to its actual use. There are plenty of questions regarding this issue on this site, but most of them either have no clear, good answer or have a solution based on checking grid positions. To restate the problem that I and many others are having, if you have a tiled surface, like a wall or a floor, consisting of several smaller component rectangles, and you traverse along them with another rectangle with force being applied into that structure, there are cases where the object gets caught on a false collision on an edge that faces the inside of the shape. I have spent a lot of time thinking about how I could possibly solve this without having to resort to a grid-based system, and I realized that physics engines do this properly. What I want to know is how they do this. What do physics engines do beyond basic SAT that allows this kind of proper collision resolution in complex environments? I've been looking through the source code to Box2D trying to find out how they do it but it's not quite as easy as looking at a Collision() method. I think I'm not good enough at physics to know what they're doing mathematically and not good enough at programming to know what they're doing programmatically. This is what I aim to fix.

    Read the article

  • Getting Steam.exe to run through a http proxy

    - by Kryptonite
    I'm sitting behind an http proxy which Steam refuses to go through. Trying Proxifier to fix the solution rendered an error about having to use an https proxy, though research shows that it shouldn't need one. Is it possible to make a target parameter in a shortcut? ie. "C:\Program Files\Steam\Steam.exe" --http-proxy=myusername:mypassword@SERVERNAME:8080 I have the server name and port number, though I'm yet to understand the relevance of 'myusername:mypassword', or infact which username and password these instructions were referring to. Of course, if a target parameter wouldn't work, would there be another way to get Steam to work?

    Read the article

  • Why do mapped drives only reappear after logging out and back in, and not after a reboot?

    - by razumny
    I work in a corporate environment, where we use mostly Windows 7 Professional computers, though some legacy applications are still being run on Windows XP. We have security in place on the network not to allow access to network resources to computers that are not members of Active Directory. When logging in, our users get their home folder and a common network drive mapped to H: and F:, respectively. Sometimes, this does not happen, and the drives are not mapped. The solution is to have the user log off, and back in to Windows. If they reboot, the drives remain unmapped. Does anyone know why this may be?

    Read the article

  • From the Coalface - 3 - Work as hard as you can to be as lazy as you can!

    - by TATWORTH
    The saga of the Change Log A recent conversation reminded me of the need for change logs within a database, to record when various change scripts were run. Creating such the required table is simple. A typical table for this consists of: Id - identity Integer primary key ChangeFileName - NVARCHAR(128) to hold the name of the file run. DateAdded - DateTime non-null with default value of getutcdate() Purpose - NVARCHAR(128) Rerunnable - Bit non-null default 0. By good design of the table only two data values normally need to be supplied. Two stored procedures, one for inserting data and one to list in reverse sequence the log complete the database essentials. The complete implementation can be found in the CommonData solution at http://CommonData.CodePlex.Com By including a call the add Change Log stored procedure, each script can log its name and purpose for posterity. The scripts that were applied to say the UAT system and their sequence of application can be readily identified for running on the Live system. Formatting XML XML is often produced as one continous string with no embedded CR/LF. To get it into human readable form, open it in visual studio, swap to another tab and back and click the format document button. The XML will then be nicely formatted!

    Read the article

  • How to make sure clients update their browser cache when my website is updated?

    - by user64204
    I am using the HTTP 1.1 Cache-Control header to implement client-side caching. Since I update my website only once a month I would like the CSS and JS files to be cached for 30 days with Cache-Control: max-age=2592000. The problem is that the 30-day period defined by Cache-Control doesn't coincide with the website update cycle, it starts from the moment the users visit the site and ends 30 days later, which means an update could occur in the meantime and users would be running with outdated content for a while, which could break the rendering of the website if for instance the HTML and CSS no longer match. How can I perform client-side caching of content for periods of several days but somehow get users to refresh their CSS/JS files after the website has been updated? One solution I could think of is that if website updates can be schedule, the max-age returned by the server could be decreased every day accordingly so that no matter when people visit the website, the end of caching period would coincide with the update of the website, but changing the server configuration every day goes against one of my sysadmin principles (once it's running, don't touch it).

    Read the article

  • OpenGL : sluggish performance in extracting texture from GPU

    - by Cyan
    I'm currently working on an algorithm which creates a texture within a render buffer. The operations are pretty complex, but for the GPU this is a simple task, done very quickly. The problem is that, after creating the texture, i would like to save it. This requires to extract it from GPU memory. For this operation, i'm using glGetTexImage(). It works, but the performance is sluggish. No, i mean even slower than that. For example, an 8MB texture (uncompressed) requires 3 seconds (yes, seconds) to be extracted. That's mind puzzling. I'm almost wondering if my graphic card is connected by a serial link... Well, anyway, i've looked around, and found some people complaining about the same, but no working solution so far. The most promising advise was to "extract data in the native format of the GPU". Which i've tried and tried, but failed so far. Edit : by moving the call to glGetTexImage() in a different place, the speed has been a bit improved for the most dramatic samples : looking again at the 8MB texture, it knows requires 500ms, instead of 3sec. It's better, but still much too slow. Smaller texture sizes were not affected by the change (typical timing remained into the 60-80ms range). Using glFinish() didn't help either. Note that, if i call glFinish() (without glGetTexImage), i'm getting a fixed 16ms result, whatever the texture size or complexity. It really looks like the timing for a frame at 60fps. The timing is measured for the full rendering + saving sequence. The call to glGetTexImage() alone does not really matter. That being said, it is this call which changes the performance. And yes, of course, as stated at the beginning, the texture is "created into the GPU", hence the need to save it.

    Read the article

  • Nginx load distribution and multi-domain SSL

    - by Steve Clark
    I'm researching into the best methods of two new parts of our infrastructure, hopefully finding a single solution for both. 1) We're currently running a single application server, and we're going to be adding an additional application server and load balance between the two. 2) We handle a few thousand domains across the application server(s), and we're looking to support SSL. The best method i've come across so far is using nginx for it's Load Distribution to serve the requests to the application servers, and for it's SSL support. If a request is using SSL, nginx accepts the request on, terminates SSL and pipes to apache (app servers). Now, that's all good, but i'm yet to figure out how we can let nginx handle multiple domains using SSL. We're potentially looking at using UCC SSL Certs, so we can support 150 domains on a single certificate, with each cert on a single IP. I'm all new to this (My experience is just with physical load balancers and a single domains on SSL), so any advice would be very much appreciated.

    Read the article

  • Problem transferring file to FTP server - 503 Could not create file

    - by blntechie
    I try to transfer a file from SSIS package to a FTP server provided by my client. When I try connecting and transfer using a FTP tool (WinSCP) manually it works fine and file was transferred. But when i use the ftp command in command prompt and try to transfer using "send" command i get an error "553 could not create file". This also happens from SSIS package. I'm not in control of the FTP server but if anyone can provide a solution I can pass it to the admin or anything to be done from my end?

    Read the article

  • Fresh install of nginx causes browser to download index.html instead of opening it

    - by 010110110101
    When I view this in Chrome, http://localhost:90 the file is downloaded instead of displayed in Chrome. This question has been asked a lot of times on SO, but about index.php files. My problem is a plain jane HTML file, not a PHP file. That hasn't been asked yet. I was hoping the solution would be similar, but I haven't been able to figure it out. Here's my example.com.conf: server { server_name localhost; listen 90; root /var/www/example.com/html index index.html location / { try_file $uri $uri/ =404; } } My index.html file contains only two words, no markup Hello World I think it's the mime.types. The mime.types file has the entry for html in it. This is a fresh nginx install. nginx -t reports "test is successful"

    Read the article

  • What is the simplest way for a slippy SVG visualization?

    - by totymedli
    I have a big SVG file representing a complicated graph with hundreds of points. I want to represent this in a web page. My idea was that I could make it like Google Maps represent their maps, in those slippy, dragable, moveable maps. I'am looking for an easy and fast JavaScript library which could do the work. What I need for my "map" is the drag/move, zoom ability, and some way to click on the points of the picture, which makes a little information apear about that point, like Google maps markers. I'am looking for a free/open source library. I saw some solutions but I'am uncertain about them, and none of them seemed to be perfet: Polymaps - I love the technique it uses, but I don't know much about this library. Leaflet - I love the simplicity of it, but I dont know how could I apply it for my SVG. Raphael - I heard the awesomeness of this, but It seemed a lots of work to do this task. What would be the best/easiest solution for my problem, and what is your opinion aboute the above libraries?

    Read the article

  • WAMP - phpmyadmin 403 error from localhost, but not from IP 127.0.0.1?

    - by kdub
    I was wondering if anyone can explain this to me. I installed WAMP 2.2. I opened up the dashboard menu, and clicked localhost. I was brought to the WAMP localhost home page. I clicked on phpmyadmin and get a 403 access is denied error message. However, if I type in the ip address 127.0.0.1/phpmyadmin in the address bar, then VOILA, I am in (however, I am not prompted to login to phpmyadmin, I am instantly brought to the phpmyadmin homepage) Please note, I have found the solution to get rid of the 403 error message when entering phpmyadmin from the the localhost extension (change the Deny all setting to Allow all); so I am not asking how to solve that, but my question is: Why if localhost and 127.0.0.1 go to the same spot, will when using the IP addres, I am granted access to subsequent applications and with localhost I am not? Any feedback would be greatly appreciated.

    Read the article

  • Any Windows based OpenID servers out there? [closed]

    - by Brian Knoblauch
    I've been looking to setup an OpenID server for a special project, but haven't found any workable OpenID server software packages. Originally was looking for a *nix solution, and found several, but they all had some kind of issue. So far I've tried JOIDS, community-id, and a couple others I unfortunately can't remember the names of. I've also come to the conclusion that even if I had managed to get one of those going that the management/upgrade cycles would have placed undue burden on the company (only a couple part time sysadmins with *nix knowledge, the day to day people are primarily Windows). So, I'm hoping that there's a Windows one out that will be functional that someone knows about and will be easy for a minimal support environment...

    Read the article

  • Stop moving page title in Firefox?

    - by lilydjwg
    Some websites I'm using always make the web title move or blink when, for example, a new mail has just arrived. It is really disturbing as it's not an emergency and usually I'm busy browsing other pages. I can't stand it any longer, but I have to use those services. I want to stop this kind of behavior either by installing a new extension or writing a Greasemonkey script. But I can't find one and have no idea about how to detect and stop the JavaScript code. What is the solution?

    Read the article

  • Non interactive git clone (ssh fingerprint prompt)

    - by qwe
    I want to clone a repo in a non-interactive way. When cloning, git asks to confirm host's fingerprint: The authenticity of host 'bitbucket.org (207.223.240.182)' can't be established. RSA key fingerprint is 97:8c:1b:f2:6f:14:6b:5c:3b:ec:aa:46:46:74:7c:40. Are you sure you want to continue connecting (yes/no)? no How do I force "yes" every time this questions pops up? I tried using yes yes | git clone ..., but it doesn't work. EDIT: Here's a solution: Can I automatically add a new host to known_hosts? (adds entires to known_hosts with ssh-keyscan).

    Read the article

  • Windows 7 / Internet Explorer 8

    - by Rene
    I am a shop owner at zazzle.com. About six weeks ago, when my computer was running on Windows XP/IE7, my sites, as well as zazzle's homepages went out on me. I can only see part of each page. Since that time, I have a new computer running Windows 7/IE8, thinking that would solve the issue. It did not. Zazzle's emails told me to download Firefox and/or download Internet Explorer 7. I tried Firefox and was getting a different problem at the zazzle site. Now I was getting only the 'view source' pages on zazzle's homepages and my own shop sites as well. Question: Can I download IE 7 onto my IE 8 computer? Can this be done without loading that compilation of internet explorer 1 through 8? What do you think is the best solution to this problem?

    Read the article

  • Url-based web site publishing on Windows Server platform

    - by Maxim V. Pavlov
    I have a Windows 2008 Enterprise SP2 server in a datacenter. It is 32bit OS. I need to be able to do a "smart" url-based web site publishing. So that with a single external IP I can publish many sites on port 80, and some firewall logic resolves, based on a requested URL, which site in IIS gets the request. Forefront TMG 2010 has this feature, but it is not supported on 32bit systems. Is there a software solution that can satissfy my need on Windows 2K8 platform? Thank you. P.S. Perhaps there is a workaround or a tweak to do what I need in IIS?

    Read the article

  • Default mount options on auto-mounted NTFS partitions (how to add `noexec` and `fmask=0111`?)

    - by jetxee
    I use auto-mounting of external USB devices, and it works as expected, except that NTFS partitions are mounted with executability flag on. For example: /dev/sdb1 on /media/Elements type fuseblk (rw,nosuid,nodev,allow_other,blksize=4096,default_permissions) All normal files are -rwxrwxrwx on this partition. I am not happy with the xs. I know I can have it mounted the way I want if I pass the fmask=0111 option. Now I use Lucid, and suppose it uses some new auto-mounting mechanism (gvfs-mount?), but I don't really know how the default mounting options can be changed now. Gconf settings in /system/storage/default_options/ntfs/mount_options have no effect. So, how do I make fmask=0111 the default automounting option for all NTFS partitions? (I'd be grateful also if someone explains how the current automounting mechanism works, how to configure it, and if the default mounting options are hard-coded, what I have to recompile to change them). I know that I can put a line in the /etc/fstab and/or mount manually, but this is not the solution I want, because 1) I don't want to edit /etc/fstab for each and every external drive I use, 2) fstab records appear in the Places pane of Nautilus, even if the drives are not present. The questions is how to change the defaults.

    Read the article

< Previous Page | 549 550 551 552 553 554 555 556 557 558 559 560  | Next Page >