Search Results

Search found 4759 results on 191 pages for 'depth buffer'.

Page 99/191 | < Previous Page | 95 96 97 98 99 100 101 102 103 104 105 106  | Next Page >

  • Webcast: Maximum Availability with Oracle GoldenGate

    - by Markus Weber
    Make sure to register here for this upcoming live webcast, with great speakers, and a live Q&A chat session.It is scheduled for Thursday, December 13th, 2012, at 10am Pacific Time. There are no doubts about the strength, and depth, of Oracle's offerings for highly available database environments. But depending on the nature of your data center environments, your different uses cases and workloads, your service level agreements and expectations, there might still be some open questions about what the best solutions and architectures are to assure high availability and/or affordable, efficient disaster recovery. Therefore, this webcast will answer questions such as: How does Oracle GoldenGate fit into Oracle's Maximum Availability Architecture ? When to use Oracle Active Data Guard ? When Oracle GoldenGate ? When both ? When Oracle GoldenGate is a definite requirement for Database High Availability ? Overall, what are new high availability features in Oracle GoldenGate 11g Release 2 ? 

    Read the article

  • What's the difference or purpose of a file format like ELF when flat binaries take up less space and can do the same thing?

    - by Sinister Clock
    I will give a better description now. In Linux driver development you need to follow a specification using an ELF file format as a finalized executable, i.e., that right there is not flat, it has headers, entry fields, and is basically carrying more weight than just a flat binary with opcodes. What is the purpose or in-depth difference of a Linux ELF file for a driver to interact with the video hardware, and, say, a bare, flat x86 16-bit binary I write that makes use of emulated graphics mode on a graphics card and writes to memory(besides the fact that the Linux driver probably is specific to making full use of the hardware and not just the emulated, backwards compatible memory accessing scheme). To sum it up, what is a difference or purpose of a binary like ELF with different headers and settings and just a flat binary with the necessary opcodes/instructions/data to do the same thing, just without any specific format? Example: Windows uses PE, Mac uses Mach-O/PEF, Linux uses ELF/FATELF, Unix uses COFF. What do any of them really mean or designate if you can just go flat, especially with a device driver which is system software.

    Read the article

  • Will the Global Demand for Water Outstrip the Supply by 2030?

    - by Evelyn Neumayr
    A recent study conducted by the Economist Intelligence Unit and sponsored by Oracle Utilities, titled “Water for All?”,considers the preparedness of utilities to supply water to the current global population of over 7 billion people, with a further 1 billion expected by 2030. It compares strategies used by utilities in 10 major countries to address this challenge. This study’s findings show that wide-ranging water management efforts and large-scale investments must be made if utilities are to meet near-certain water stress—demand outstripping supply—by 2030. The report is based on an online survey of 244 executives of water utilities in these countries, supplemented by in-depth interviews with 20 water utility executives and independent experts. The research concludes that utilities worldwide expect to meet future demand, despite increased supply pressure on supplies, due to improvements in water productivity that the wide range of measures utilities and governments will take to ensure that water is used more efficiently. Read more about this here.

    Read the article

  • I am new to game development, what do I need to know? [closed]

    - by farmdve
    I am unsure if this question is a duplicate, I hope it isn't. Are there any resources on the terminology when doing game development? Because, even if you tell me to learn some graphics API, how would I understand the things it does, if I am not well into the terminology(voxel,mesh,polygon,shading). What about the math that is involved in the game(geometry) or the concept of the gravity,collision detection in the game and their respective maths? I am very bad at math, never was good, because I have ADHD, but I won't give up just yet. I look at a game, and I see "textures", but how am I walking on them, how do they take substance so I don't fall off of them? And depth? This is what I need information about, not just a link to a library like SDL(which I have compiled under MinGW and MinGW-W64) and tell me to learn it and the cliché answer "start simple/small". I hope the question(s) are not too vague.

    Read the article

  • How to get the Dash and HUD to appear. (and stop Unity spewing error messages.)

    - by Ubuntiac
    I just installed Ubuntu 12.04 on my wifes Dell Inspiron 1501, which uses an R300 ATI graphics chip. Neither the Dash or HUD appear when pushing the appropriate key. When I try unity --reset & in the terminal, I see that over and over it's spitting out: r300: CS space validation failed. (not enough memory?) Skipping rendering. This is just after starting Ubuntu with no apps open, so I find it hard to believe that just rendering the Dash / HUD is completely blowing out the VRAM. Any suggestions on getting this working? /usr/lib/nux/unity_support_test -p shows OpenGL vendor string: X.Org R300 Project OpenGL renderer string: Gallium 0.4 on ATI RS480 OpenGL version string: 2.1 Mesa 8.0.2 Not software rendered: yes Not blacklisted: yes GLX fbconfig: yes GLX texture from pixmap: yes GL npot or rect textures: yes GL vertex program: yes GL fragment program: yes GL vertex buffer object: yes GL framebuffer object: yes GL version is 1.4+: yes Unity 3D supported: yes All sections say "YES"

    Read the article

  • SQLServerCentral Webinar Series #21 - Forgotten Rings & Other Monitoring Stories

    Most common monitoring metrics are important and useful, especially over time, but they can fall short. How do you gather information to determine, for example, if you have buffer cache pressure? Register now for the free webinar. Wednesday, October 17 2012 4:00pm - 5:00pm BST Are you sure you can restore your backups? Run full restore + DBCC CHECKDB quickly and easily with SQL Backup Pro's new automated verification. Check for corruption and prepare for when disaster strikes. Try it now.

    Read the article

  • What is a right datatype in C++ for OpenGL scene representation with use of GLSL

    - by Rarach
    I am programming in C++ OpenGl with GLSL. Until now I have been using a data structure that is composed of std::vector filled with structures of vertexes and with their parameters (position , normal, color ...) as a global variable for all the code. My question is, as I am using VBOs for drawing - is this a good approach to this problem? I am asking because I happen to have a lot of memory related trouble with this structure. I am trying to generate a terrain with a lot of vertices - more than 1 million. This seems to work, but as I refill the buffer I get memory related issues (crushes that occur, more or less randomly). So again the question is, is this a good data structure to use / and look for the faults in my code, or should I change to something else? Or what data structure would be advisable?

    Read the article

  • Branding Changes for Java EE6

    - by Paul Sorensen
    Hi Everyone, As we move the Java EE6 exams from beta to production, you may notice that we have made a slight change in the branding. Instead of being branded Oracle Certified Professional (OCP), these new credentials are now branded Oracle Certified Expert (OCE). One area where we use the Expert brand is for credentials where the technology is advanced or broader than the path based credential requires. Some are high-end add-on certifications, and others have significant additional technological breadth. In these cases, the Expert brand is an indication that someone is tested in more advanced or in-depth skills - beyond the traditional path-based certification. A few examples are RAC Expert for DBAs, or SQL Expert - also for DBAs. Because (1) all of the Java EE6 credentials require that candidates become certified first in Java SE6, and (2) many people earn more than one Java EE credential, we felt that the Expert branding would be more appropriate. Thanks,

    Read the article

  • Slow writing HDD speed, Ubuntu 12.04 64-bit, Thinkpad T520i

    - by pyc
    It seems that (but I'm not completely sure), that when I'm copying files from gigabit network to HDD, I can't use full potential of the network which in my case is about 60 MB/s, because HDD writing is so slow like lower than 10 MB/s, and also it's slowing down the whole system which becomes pretty much unresponsive, almost impossible to work with. Copying files to samba share residing at Ubuntu machine, connected to share from Windows 7, I'm completely sure my network equipment is OK, and there's no CPU intensive process on Ubuntu except smbd getting about 10-20% from time to time which I think is OK. Something here is burried deep I think, maybe even in kernel. Already tried to switch from AHCI to compatibility mode, and turning acpi on and off - nothing helped. So it's like HDD buffer is full and emptying slowly while machine is sluggish, load is about 3 to 4. Somebody experienced the similar problems? Some help on troubleshooting process and identifying the cause would be helpful too :) Thanks!

    Read the article

  • Why fgetc too slow?

    - by user14284
    I've written a program that reads an whole file via fgetc: while ((c = fgetc(f)) != EOF) { ... } But the program is too slow. When I changed fgetc to fread, static unsigned char buf[4096]; while ((n = fread(buf, 1, sizeof(buf), f)) > 0) { ... } the program works about 10 times faster. Why? As I know, fgetc is a buffered function, so it should work as fast as the second version with explicit buffer, isn't it?

    Read the article

  • Make public webcam. Which protocol, which codec. (Using VLC)

    - by gsedej
    Hi! I want to use my old (1GHz) PC as webcam video stream server (like you can see those road cameras). I thought of using VLC and already tried using http output but it was not really good. Too cpu hungry, too big stream (kBps), not stable... I been reading VLC how-to's but thre is still a question. Which output should I use? Http, RTSP, UDP? I want to make for more than one computer at the same time (multicast). Which codec should be good? PC is not so fast so it shouldn't be too cpu hungry codec. Mpeg2, mpeg4, xvid? how much video buffer should I use (vb=?)? What about setting IP and ports? So I need some help with ideas, but if someone can make a VLC command line it's even better :) Oh, computer has direct internet connection and own IP.

    Read the article

  • are there compiler options in clang? [on hold]

    - by Deohboeh
    I am learning from The C++ Primer. One of the exercises is to compile a program with arguments in main(). For this I am trying to use mac terminal. I need to compile a C++11 Unix executable file named "main" which takes “f" as an argument. I am using Xcode 4.6.3 on OS X Lion. I compiled the program with clang++ -std=c++11 -stdlib=libc++ main.cpp -o main. But don’t know what to do next. I found -frecord-gcc-switches while searching compiler options on google. It does what I need to do. Is there a clang version of this? Please use simple language. I have never used command line before. I tried going through the clang user guide but a lot of it is out of my depth.

    Read the article

  • how to fix error with ati driver ubuntu 11.10 X Error of failed request?

    - by user34785
    i install driver by this link : http://cisight.com/install-amd-radeon-hd-6470m-and-solve-overheat-on-ubuntu-1110-oneiric/ but will time i run fglrxinfo , display it : X Error of failed request: BadRequest (invalid request code or no such operation) Major opcode of failed request: 136 (GLX) Minor opcode of failed request: 19 (X_GLXQueryServerString) Serial number of failed request: 12 Current serial number in output stream: 12 my xorg is : Section "ServerLayout" Identifier "aticonfig Layout" Screen 0 "aticonfig-Screen[0]-0" 0 0 EndSection Section "Module" Load "fglrx" EndSection Section "Monitor" Identifier "aticonfig-Monitor[0]-0" Option "VendorName" "ATI Proprietary Driver" Option "ModelName" "Generic Autodetecting Monitor" Option "DPMS" "true" EndSection Section "Screen" Identifier "aticonfig-Screen[0]-0" Device "aticonfig-Device[0]-0" Monitor "aticonfig-Monitor[0]-0" DefaultDepth 24 SubSection "Display" Viewport 0 0 Depth 24 EndSubSection EndSection how may i solve it? thanks

    Read the article

  • Unity3D 3.5 pro - Moving the camera vs setting draw distance

    - by stoicfury
    I move the camera mostly via right-click + WASD, sometimes with [shift] if I want it to move faster. Occasionally, instead of moving my camera, it alters the draw distance / FOV / some visual aspect of the editing scene that causes trees and other object to disappear when I scroll enough, and eventually even the terrain starts disappearing. It is not m "zooming out". My camera does not move, the width and height of the FOV stays the same (one might say the depth is being altered though). What key am I hitting to cause this to happen, and is it possible to disable it? side note: "keybinds" is probably the most spot-on tag for this question but it doesn't exist (surprisingly) and I lack the rep to create it.

    Read the article

  • Extracting, Transforming, and Loading (ETL) Process

    The process of Extracting, Transforming, and Loading data in to a data warehouse is called Extract Transform Load (ETL) process.  This process can be used to obtain, analyze, and clean data from various data sources so that it can be stored in a uniform manner within a data warehouse. This data can then be used by various business intelligence processes to provide an organization with more of an in depth analysis of the current state of the company and where it is heading. A standard ETL process that might be used by a health care system may include importing all of their patients names, diagnoses and prescriptions in to a unified data warehouse so that trends can be spotted in regards to outbreaks like the flu and also predict potential illness that a patient might be affected by based on other patients with similar symptoms.

    Read the article

  • XNA VertexBuffer.SetData performance suggestions

    - by CodeSpeaker
    I have a 3d world in a grid layout where each grid cell contains its separate vertex and index buffer for the mesh/terrain of that cell. When the player moves outside the boundaries of his cell, i dynamically load more cells in his walking direction based on his viewing distance. This triggers x number of vertex and indexbuffer initializations depending on how many cells that needs to be generated and causes the framerate to drop annoyingly during this time. The generation of terrain data is handled in a separate thread and runs smoothly. The vertex and index buffers are added during the update cycle of the game loop. I´ve tried batching the number of cells to be processed to avoid sending too much data at once into the buffers, which worked ok at a shorter viewing distance (about 9 cells to process), but not as well at greater distances with around 30 cells to process. Any idea how i can optimize this?

    Read the article

  • Is it possible to save output from a command to a file subsequent the command already has been executed?

    - by NES
    Does an elegant way exists to save the output of a command subsequent to a file as long as the terminal window is open? I mean once the command has been executed in the terminal. The output is still there in terminal. Now i could copy & paste all the lines and save it to a file. But perhaps does a method exist to somehow write the output buffer of a terminal window to a file or even better the output of an already executed command?

    Read the article

  • What file formats and conventions should I support to make my game engine artist-friendly?

    - by Avi
    I'm writing a game engine, and I want to know what I should do to make it more artist-friendly. I don't want to be too limiting in terms of what file formats I support, etc. Some specific questions: Are there specific formats artists like to model in? Does it not matter because the 3D modeler abstracts the data storage away? Is it okay if I don't support per-vertex coloration in my game engine? If I have to store a diffuse, specular, ambient, and emissive color value for each vertex, it doubles the size of vertices in the buffer. Is it reasonable to ask artists to do all these things in textures / maps? Any other tips you have about making it so that artists have to adapt their style to my specific engine as little as possible would be nice.

    Read the article

  • OpenGL ES 2.0. Sprite Sheet Animation

    - by Project Dumbo Dev
    I've found a bunch of tutorials on how to make this work on Open GL 1 & 1.1 but I can't find it for 2.0. I would work it out by loading the texture and use a matrix on the vertex shader to move through the sprite sheet. I'm looking for the most efficient way to do it. I've read that when you do the thing I'm proposing you are constantly changing the VBO's and that that is not good. Edit: Been doing some research myself. Came upon this two Updating Texture and referring to the one before PBO's. I can't use PBO's since i'm using ES version of OpenGL so I suppose the best way is to make FBO's but, what I still don't get, is if I should create a Sprite atlas/batch and make a FBO/loadtexture for each frame of if I should load every frame into the buffer and change just de texture directions.

    Read the article

  • Bridging the gap between learning language fundamentals and actually making useful software?

    - by Anonymous -
    I'm learning C# via the "Essential C#" Lynda.com video tutorials and plan to read a couple of books that cover things in more depth afterwards. My question is where I should head to learn more after that? I've done things like project Euler in the past, but I find they don't really help me learn anything other than basic program control flow and features. I've looked at many open-source projects but pretty much everything still looks overwhelmingly complicated at this stage. What would you recommend I look at to help me build useful applications that are a bit beyond the millions of console applications I must've written thus-far? Should I be looking at books specifically on learning/working with the .NET framework, or just biting my lip and continue working through open source projects until they start to make sense?

    Read the article

  • Looking for parallel programming problem

    - by Chris Lieb
    I am trying to come up with a problem that is easily solvable in a parallel manner and that requires communication between threads for a test. I also am trying to avoid problems that require require random waits, which rules out dining philosophers and producer-consumer (bounded buffer), two of the classics. My goal is for the student to be able to write the program in less than 20-30 minutes in front of a computer not knowing of the problem beforehand. (This is to prevent preparation more than to come up with something novel.) I am trying to stress the communication aspect of the program, though the multi-threaded nature is also important. Does anyone have some ideas? Edit: I'm using Google Go for the language and testing comprehension of the goroutines/channels combo vs an actors library that I authored.

    Read the article

  • Is programming a SubCulture? [closed]

    - by Trufa
    I was going through this article: http://en.wikipedia.org/wiki/Subculture Which got mee thinking is programming a subculture? After the a while I started thinking it really hard, and if you go really in depth this is a very complex and interesting question to ask. YOu can even ask yourself if (heavy) internet (social) users are an subculture and programmers a culture within. I think it might be an interesting discussion, hope you like it! NOTE: I linked the wiki article because it might be a good baseline, maybe you can base you answer on Ken Gelder´s proposal to distinguish subcultures. But it should be based on a little bit more that intuition. Thanks in advance! Trufa

    Read the article

  • How can I increase resolution when impossible to set up nvidia driver?

    - by Asimov4
    I was stuck at 1024x758 before creating an /etc/X11/xorg.conf file as follows which works well to allow me to have a resolution of 1920x1200 with the vesa driver. Section "Device" Identifier "Configured Video Device" Driver "vesa" EndSection Section "Monitor" Identifier "Configured Monitor" HorizSync 24.0 - 83.0 VertRefresh 50.0 - 75.0 EndSection Section "Screen" Identifier "Default Screen" Monitor "Configured Monitor" Device "Configured Video Device" SubSection "Display" Viewport 0 0 Depth 24 Modes "1920x1200" EndSubSection EndSection But now my display is kind of slow. I have an NVIDIA GeForce 8600GT but really cannot find how to set up the driver to work. Are there any alternatives to VESA that would work faster?

    Read the article

  • Gone in 60 Seconds: An Insecure Database is an Easy Target

    - by Troy Kitch
    According to the recent Verizon Data Breach Investigations Report, 98% of breached data originates from database servers and nearly half are compromised in less than a minute! Almost all victims are not even aware of a breach until a third party notifies them and nearly all breaches could have been avoided through the use of basic controls. Join us for this November 28th webcast to learn more about the evolving threats to databases that have resulted in over 1 billion stolen records. Also, hear how organizations can mitigate risks by adopting a defense-in-depth strategy that focuses on basic controls to secure data at the source - the database. There's no turning back the clock on stolen data, but you can put in place controls to ensure your organization won't be the next headline. Note, this webcast will be recorded for on-demand access after November 28th. 

    Read the article

  • Algorithm for Learning development

    - by user9057
    Hi all, This is a fairly general question. I know a bit of Perl and Python and I am looking to learn programming in more depth so that once I get the hang of it I can start developing applications and then websites. I would like to know of an algorithm (sequence of steps :)) that could describe my approach towards learning programming in general. I have posted small questions on Perl/Python and I have recieved great help from everyone. Note:- I am not in a hurry to learn. I know it takes time and that's fine. Please give any suggestions you think are valid. Also, please don't push me to learn Lisp, Haskell etc - I am a beginner.

    Read the article

< Previous Page | 95 96 97 98 99 100 101 102 103 104 105 106  | Next Page >