Search Results

Search found 1603 results on 65 pages for 'coordinate transformation'.

Page 30/65 | < Previous Page | 26 27 28 29 30 31 32 33 34 35 36 37  | Next Page >

  • Centralizing Chart of Accounts Management Across Oracle ERP and EPM Applications with Oracle Hyperion Data Relationship Management

    Most enterprises today have multiple GL/ERP systems - each with their own set of accounts, structures and systems for financial and management reporting. Mergers, acquisitions and reorganizations inject constant change into the process - through new accounts, entities, and locations. Accommodating an organization's unique view of the business while still maintaining accurate collection, measurement and reporting at the corporate level makes synchronization of chart of accounts across multiple systems a challenge. In this podcast, you'll hear about how Oracle Hyperion Data Relationship Management allows you centralize and align different financial perspectives into your corporate reporting standards. This end-user oriented, technology agnostic hierarchy management solution enables organizations to coordinate the management of chart of accounts across the enterprise and save a significant amount of time and effort.

    Read the article

  • Where in a typical rendering pipeline does visibility and shading occur?

    - by user29163
    I am taking a computer graphics course. The book and the lecture notes are vague on the on the order of flow between the different steps in the rendering process. For example, if we have specified a view in a scene, and then want to perform a projection transformation for that given view, then we have to go through a sequence of transformations. In the end we end up with a normalized "viewcube" ready to be mapped 2D after clipping. But why do we end up with a cube (ie 3D thing), when a projection results in projecting the 3D objects to 2D. (depth information is lost?) The other line of reasoning is that all information further needed is stored within the "cube" and that visibility detection and shading is performed with respect to this cube and then we perform rasterezation.

    Read the article

  • cVidya’s MoneyMap Achieves Oracle Exadata Optimized Status

    - by Javier Puerta
    cVidya's MoneyMap running on Oracle Exadata provides extreme performance, including 4x-16x improvement in high data load rates, 4x faster data transformation and reconciliation, and query speeds - from a 2.5 billion record index –  improved from hours to few seconds! The MoneyMap solution enables operators to reconcile information from all network, operations and business support systems and through an on-going automated process, it detects problem areas which impact profitability as a result of revenue leakage, data inconsistencies or resources that are not being used efficiently. Once detected, MoneyMap provides tools to promptly correct and manage the problems to achieve profit maximization Learn more here.

    Read the article

  • 2d engine with WebGL

    - by Philipp
    I am currently developing a HTML5 canvas game with a sprite-based 2d engine. The engine is working well so far, but now I thought that I could maybe add some cool graphic effects when I would re-implement the whole engine with WebGL. The whole graphic engine would stay strictly two-dimensional (think of it as a 3d world where the z coordinate of all vertices is 0.0). What I hope to gain from this is the ability to use pixel shaders to create special effects. I also think that the performance could improve. What do you think about that plan?

    Read the article

  • Google se lance dans les solutions de géolocalisation en temps réel pour les entreprises, mais uniquement sur Android

    Google se lance dans les solutions de géolocalisation en temps réel Pour les entreprises, mais uniquement sur Android « Imaginez : vous êtes un opérateur pour un fournisseur d'électricité. Vous recevez un appel concernant une panne électrique et vous devez rapidement dépêcher sur place un de vos techniciens. Google Maps Coordinate peut vous aider ». Voici comment Google présente son nouveau service aux entreprises déclinés des Google Maps. Cette solution à la croisée de la géolocalisation en temps réel et des smartphones sort au moment où cabinet d'études IDC estime qu'il y aura plus de 1,3 milliard de travailleurs mobiles d'ici 2015 (soit 37,2% des employés au total).

    Read the article

  • Game Database Connectivity Java

    - by The Kraken
    I'm developing a simple multi-player puzzle game in Java. Both players should be able to view the same game board on his own computer. Then, when one player makes an action in the game (ex. drags an object onto a coordinate space), the game's view should update automatically on the other computer's game screen. I'd like all this to happen over the internet, not requiring both computers to be on the same LAN connection. If I need to use SQL/PHP to accomplish this, I'm unsure how to design the database to accomplish something as simple as the following: Player A drags element onscreen Game sends coordinates of element to database/server Player B's computer detects a change to an item in the database Player B's computer grabs the coordinates of Player A's item Player B's machine draws onscreen elements at the received coordinates Could somebody point me in the right direction?

    Read the article

  • Mirth: Transforming a response message in a separate Channel and returning it to original channel

    - by Ryan H
    I have a channel that takes HL7v2 message and converts it to HL7v3. It invokes a SOAP web service and receives a response in HL7v3. I need to convert that response back to HL7v2. Currently: I "Send Response to:" my second channel. That can convert it fine back to HL7v2, but it doesn't seem to return a response message. I want that second transformation to be the response to the initiator which is an LLP Listener.

    Read the article

  • Image classification using openCV and weka

    - by simk
    Hi i want to do image classification, so i am planning to use openCV for the preprocessing of image and weka to check which ML algorithm gives best result, so the problem i am facing is converting the image data in to weka ARFF file format, when i apply some image transformation to image and write the image data it become so large and not sure how to define @ATTRIBUTE for that. If you have done similar thing before please suggest how can i solve this and it particularly doesn't have to be openCV i can use other tools also, please suggest.

    Read the article

  • WPF: How to autosize Path to its container?

    - by 742
    I have a Path that must resize to its StackPanel container. <StackPanel x:Name="TrackSurface"> <Path Fill="AliceBlue" Stroke="Black" StrokeThickness="1" Data="{StaticResource TranslateZ}"> </Path> </StackPanel> I think about using a transformation bound to the container but don't know how to it actually. Can someone give me hint?

    Read the article

  • How to override filter in android's ArrayAdapter?

    - by Mike
    I have an ArrayAdapter wrapped around an ArrayList of custom objects. I'd like to write a custom filter for that adapter so that when I call getListAdapter().getFilter().filter("abc") the list will get filtered by an arbitrary transformation of "abc". I thought I would just try to override ArrayAdapter.getFilter(), but that requires I re-implement the private ArrayAdapter.ArrayFilter which requires access to a bunch of ArrayAdapter's private instances. What's the simplest way to do this?

    Read the article

  • WPF Zoom with ScaleTransform overlap all Window

    - by Andrew
    Hi, i want to create Pan & Zoom control Similar thread, so i transform control with ScaleTransform and it overlaps all window, i have ot do it with RenderTransform, becouse with LayoutTransform Pan implemented in example doesn't work. Are there any properties or templates with which i can implement behavior like this: if content of container undergo transformation (or just move), if content doesn't fit, container shows only fitted part, something like this: <Container Height="100" Width="100" DisplayOption="CutAllThatNotFit"> <Content/> </Container>

    Read the article

  • iPhone SDK Zoom and refresh PDF with Quartz

    - by Ben
    Looking at the QuartzDemo sample application, I love the speed of the PDF rending using quartz alone (that is, without using uiwebview). However, when I'm zooming in the PDF it doesn't seem to become more clear like it does in PDF view. Is there something that I can change to have the same effect when zooming in and out using multitouch? like manipulate the PDF transformation matrix or something? Thanks a bunch. --Ben

    Read the article

  • Hibernate ResultTransformer with JPA API

    - by Timo Westkämper
    Has anyone figured out a smart way to do query result transformation through a similar mechanism like specifying a ResultTransformer in Hibernate? All I can think of is transforming each result row after it has been returned by the Query. Is there any other way? For constructor projections (e.g. new DTO(arg1, arg2)) it can be defined in the JPQL query, at least for Hibernate, but how about other cases?

    Read the article

  • Drawing continuously in drawing application

    - by user146780
    I was wondering how drawing applications draw the entire time the mouse is down without having empty gaps. What I mean is, for example if the program only drew circles at the mouse's X, y coordinate, then if the mouse went too quicly it would seem like a bunch of little circles rather than a nice continuous line. How can this be done without constantly drawing a short straight line between where the mouse was 0.001 seconds ago and where the mouse now is. Thanks

    Read the article

  • msbuild TransformWebConfig task - doesn't work for App.configs?

    - by Jeff D
    I have a windows service that will need to have the same kind of transformations that the web.configs would use, but VS 2010 doesn't seem to support that. I've tried manually adding the App.Release.Config files, and then using msbuil [PROJ] /T:TransformWebConfig /p:Configuration=Release but no transformation is performed. I got a TransformWebConfig folder createed in my obj subdirectory, but that's it. Is this thing hardcoded to only work with web.configs?

    Read the article

  • Crop the image using Jquery in C# and ASP.Net

    - by Vara Prasad.M
    I am having a image in one page and i have to crop the image to the predefined borders to crop this should happen using jquery and it has to be done in C# and ASP.Net and this is done by using Client side not the server side. but the X and Y coordinate values must be saved in hidden fields and used for backend plz reply me ASAP it is very urgent ThanQ, Vara Prasad.M

    Read the article

  • SSIS how to split a single record in to two different records?

    - by Dr. Zim
    I have a Product record that has multiple "zoned" prices, one for each store that sells the product. ProductID int Name string PriceA money PriceB money PriceC money In SQL Server Integration Services, I need to split this in to multiple records: ProductID int Version string // A, B, or C Price money // PriceA if A, PriceB if B, etc. This would be within a Data Flow, I presume as a Transformation between Excel source and OLE DB destination. (Assuming OLE DB is a good destination for MS SQL server).

    Read the article

  • The Skyline Problem.

    - by zeroDivisible
    I just came across this little problem on UVA's Online Judge and thought, that it may be a good candidate for a little code-golf. The problem: You are to design a program to assist an architect in drawing the skyline of a city given the locations of the buildings in the city. To make the problem tractable, all buildings are rectangular in shape and they share a common bottom (the city they are built in is very flat). The city is also viewed as two-dimensional. A building is specified by an ordered triple (Li, Hi, Ri) where Li and Ri are left and right coordinates, respectively, of building i and Hi is the height of the building. In the diagram below buildings are shown on the left with triples (1,11,5), (2,6,7), (3,13,9), (12,7,16), (14,3,25), (19,18,22), (23,13,29), (24,4,28) and the skyline, shown on the right, is represented by the sequence: 1, 11, 3, 13, 9, 0, 12, 7, 16, 3, 19, 18, 22, 3, 23, 13, 29, 0 The output should consist of the vector that describes the skyline as shown in the example above. In the skyline vector (v1, v2, v3, ... vn) , the vi such that i is an even number represent a horizontal line (height). The vi such that i is an odd number represent a vertical line (x-coordinate). The skyline vector should represent the "path" taken, for example, by a bug starting at the minimum x-coordinate and traveling horizontally and vertically over all the lines that define the skyline. Thus the last entry in the skyline vector will be a 0. The coordinates must be separated by a blank space. If I will not count declaration of provided (test) buildings and including all spaces and tab characters, my solution, in Python, is 223 characters long. Here is the condensed version: B=[[1,11,5],[2,6,7],[3,13,9],[12,7,16],[14,3,25],[19,18,22],[23,13,29],[24,4,28]] # Solution. R=range v=[0 for e in R(max([y[2] for y in B])+1)] for b in B: for x in R(b[0], b[2]): if b[1]>v[x]: v[x]=b[1] p=1 k=0 for x in R(len(v)): V=v[x] if p and V==0: continue elif V!=k: p=0 print "%s %s" % (str(x), str(V)), k=V I think that I didn't made any mistake but if so - feel free to criticize me. EDIT I don't have much reputation, so I will pay only 100 for a bounty - I am curious, if anyone could try to solve this in less than .. lets say, 80 characters. Solution posted by cobbal is 101 characters long and currently it is the best one. ANOTHER EDIT I thought, that 80 characters is a sick limit for this kind of problem. cobbal, with his 46 character solution totaly amazed me - though I must admit, that I spent some time reading his explanation before I partially understood what he had written.

    Read the article

  • Use JTS topology in JavaFx

    - by borovsky
    I have some polygon in jts topology library. if I want to draw on javafx pane I do: Polygon poly=new Polygon();//javafx //g is geometry of jts for (Coordinate coord : g.getCoordinates()) { poly.getPoints().addAll(coord.x, coord.y); } and got extra four edges that are not expected: but the same data look good in test builder: what is wrong? order of traversing the geometry? any ideas?

    Read the article

  • Palm Mojo: Object properties

    - by tom
    There was another question with an answer saying that you can use event.down.x to get the x-coordinate of a tap event. It also said that this was not well documented in the Palm Mojo SDK. My question is how can I find the properties that I can use for any given Mojo object?

    Read the article

  • Forward geocoding from the iPhone

    - by Zap
    Hi, Given that the mapkit doesn't provide forward geocoding functionality today, can anyone provide help on how I can i use a search bar to return a latitude and longitude coordinate from a user inputted address today. If someone can provide sample code that would be great. Zap

    Read the article

  • Ray-Box Intersection during Scene traversal with matrix transforms

    - by Myx
    Hello: There are a few ways that I'm testing my ray-box intersections: Using the ComputeIntersectionBox(...) method, that takes a ray and a box as arguments and computes the closest intersection of the ray and the box. This method works by forming a plane with each of the faces of the box and finding an intersection with each of the planes. Once an intersection is found, a check is made whether or not the point is on the surface of the box by checking that the intersection point is between the corner points. When I look at rays after running this algorithm on two different boxes, I obtain the correct intersections. Using ComputeIntersectionScene(...) method without using the matrix transformations on a scene that has two spheres, a dodecahedron (a triangular mesh), and two boxes. ComputeIntersectionScene(...) recursively traverses all of the nodes of the scene graph and computes the closest intersection with the given ray. This test in particular does not apply any transformations that parent nodes may have that also need to be applied to their children. With this test, I also obtain the correct intersections. Using ComputeIntersectionScene(...) method WITH the matrix transformations. This test works like the one above except that before finding an intersection between the ray and a node in the scene, the ray is transformed into the node's coordinate frame using the inverse of the node's transformation matrix and after the intersection has been computed, this intersection is transformed back into the world coordinates by applying the transformation matrix to the intersection point. When testing with the third method on the same scene file as described in 2, testing with 4 rays (thus one ray intersects the one sphere, one ray the the other sphere, one ray one box, and one ray the other box), only the two spheres get intersected and the two boxes do not get intersections. When I debug looking into my ComputeIntersectionBox(...) method, it actually tells me that the ray intersects every plane on the box but each intersection point does not lie on the box. This seems to be strange behavior, since when using test 2 without transformations, I obtain the correct box intersections (thus, I believe my ray-box intersection to be correct) and when using test 3 WITH transformations, I obtain the correct sphere intersections (thus, I believe my transformed ray should be OK). Any suggestions where I could be going wrong? Thank you in advance.

    Read the article

< Previous Page | 26 27 28 29 30 31 32 33 34 35 36 37  | Next Page >