Search Results

Search found 29938 results on 1198 pages for 'version hunter'.

Page 521/1198 | < Previous Page | 517 518 519 520 521 522 523 524 525 526 527 528  | Next Page >

  • Flashplayer 11.1 on Ubuntu 12.04

    - by Courtney
    I am new to Ubuntu and I've visited pretty much every webpage that exists on the internet about how to install canonical Flashplayers. Ubuntu 12.04 comes with Adobe Flashplayer 11.2 but that is not the most recent version and Adobe no longer supports Ubuntu apparently. I have no idea how to install Flashplayer 11.1 via Terminal. I was able to find a site with canonicals (https://launchpad.net/ubuntu/+source/flashplugin-nonfree) but I have no clue where to go from there, though I've tried for approx. 4 hours to figure it out on my own. I am hoping someone could take me step by step through the exact code. For example, "download, then install file" is not specific enough since I am completely new to this. Please help!

    Read the article

  • 12.04 - Connecting Acer X203H monitor to new Dell XPS 15z laptop

    - by Lucy Dixon
    I have installed the latest version of Ubuntu (12.04) on my boyfriend's new Dell XPS 15z laptop. He uses a Microsoft wireless keyboard and mouse, and an Acer X203H monitor with his set-up. No problems with the keyboard or mouse, or with connecting the HP printer, but we just can't get the laptop to talk to the Acer monitor. With his old setup he used a VGA cable to connect machine & monitor. New laptop has no VGA port, but we've bought a VGA to HDMI adaptor to connect to the laptop. Have tried using Fn F2 to change the display from laptop to monitor, but it doesn't see the monitor at all. HELP! Is there a driver I can install from somewhere? Or how can I tell Ubuntu to look for the monitor on another port? Completely in the dark, and about to get in trouble!! Thanks

    Read the article

  • libnotfy1(>= 0.4.4) and libnotify1-gtk2.10 for Ubuntu 11.10

    - by Aamir
    I have recently downloaded the latest version of Lotus Symphony 3.0.1 but whenever I try to install it, it shows the following dependencies. Can anybody tell where I can find these dependencies. dpkg: dependency problems prevent configuration of symphony: symphony depends on **libnotify1 (>= 0.4.4)**; however: Package libnotify1 is not installed. symphony depends on **libnotify1-gtk2.10**; however: Package libnotify1-gtk2.10 is not installed. dpkg: error processing symphony (--install): dependency problems - leaving unconfigured Errors were encountered while processing: symphony

    Read the article

  • How do I get adobe flash working in Chrome or Chromium on Ringtail?

    - by Matt H
    I have recently upgraded to Ringtail and for some reason flash isn't working on either Chrome or Chromium. It works in firefox but I prefer Chrome as my browser so switching to firefox when a website containing flash appears is a bit annoying. According to just about every source, flash it built into Chrome and should just work even on Ubuntu. I tried removing chrome and reinstalling it, but the problem persists. I've checked about://plugins and flash is enabled. Yet when you visit http://www.adobe.com/shockwave/welcome/ What I see is "No plugin available to display this content". about://version shows Google Chrome 28.0.1500.52 (Official Build 207119) OS Linux Blink 537.36 (@152651) JavaScript V8 3.18.5.8 Flash 11.7.700.203 User Agent Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/28.0.1500.52 Safari/537.36 Command Line /opt/google/chrome/google-chrome --blacklist-accelerated-compositing --flag-switches-begin --enable-sync-favicons --sync-keystore-encryption --flag-switches-end Executable Path /opt/google/chrome/google-chrome How do I get flash working in Chrome or Chromium?

    Read the article

  • Accenture Launches Smart Grid Data Management Platform

    - by caroline.yu
    Accenture announced today it has launched the Accenture Intelligent Network Data Enterprise (INDE), a data management platform to help utilities design, deploy and manage smart grids. INDE's functionality can be enabled by an array of third party technologies. In addition, Accenture plans to offer utilities the option of implementing the INDE solution based on a pre-configured suite of Oracle technologies. The Oracle-based version of INDE will accelerate the design of smart grids and help reduce the costs and risks associated with smart grid implementation. Stephan Scholl, Senior Vice President and General Manager of Oracle Utilities said, "Oracle and Accenture share a common vision of how the smart grid will enable more efficient energy choices for utilities and their customers. Our combined expertise in delivering mission-critical smart grid applications, security, data management and systems integration can help accelerate utilities toward a more intelligent network now and as future needs arise." For the full press release, click here.

    Read the article

  • Ops Center zip documentation

    - by Owen Allen
    If you're operating in a dark site, or are otherwise without easy access to the internet, it can be tricky to get access to the docs. The readme comes along with the product, but that's not exactly the same as the whole doc library. Well, we've put a zip file with the whole doc library contents up on the main doc page. So, if you are in a site without internet access, you can get the zip, extract it, and have a portable version of the site, including the pdf and html versions of all of the docs.

    Read the article

  • Technologie Roadmap: Portlet JSR286 vs Widget/Gadget

    - by Aerosteak
    Hello IBM got me confused (again). For many years IBM have been pushing for Portlet Containers with the JSR 168 and later the JSR 286 Specification. 2008-2009, IBM the Lotus division introduced the iWidget Specification. Based on my reading, it is a more dynamic and lightweight version of the Portlets, close to Google Gadget. It uses a different paradigm than Porlet while providing the same features. A major differentiator with this kind of client side technologies is that you don’t need a big and costly Portal infrastructure. To not fall in the ‘It depends on needs’ discussions, let consider the following: * New company, no legacy portlet, no portal in place. What are your thoughts on this?

    Read the article

  • Why eclipse is hanging while in debug mode ?

    - by Pratik
    We are developing our web application using JAVA GWT framework. We are using Eclipse Indigo as a development GUI. We are facing problems while debugging the JAVA gwt application in eclipse. Most of the time, Eclipse hangs while debugging. We tried to increase the memory buffer size in eclipse but no luck. We had tried to run the eclipse in various environment like Windows, Fedora 16, Cent OS. but some how not getting positive results. Can anyone help me out to decide which OS, and eclipse or version should we have to use so can able to resolve the hanging issue? Thanks in advance. Pratik

    Read the article

  • Dynamically loading Assemblies to reduce Runtime Depencies

    - by Rick Strahl
    I've been working on a request to the West Wind Application Configuration library to add JSON support. The config library is a very easy to use code-first approach to configuration: You create a class that holds the configuration data that inherits from a base configuration class, and then assign a persistence provider at runtime that determines where and how the configuration data is store. Currently the library supports .NET Configuration stores (web.config/app.config), XML files, SQL records and string storage.About once a week somebody asks me about JSON support and I've deflected this question for the longest time because frankly I think that JSON as a configuration store doesn't really buy a heck of a lot over XML. Both formats require the user to perform some fixup of the plain configuration data - in XML into XML tags, with JSON using JSON delimiters for properties and property formatting rules. Sure JSON is a little less verbose and maybe a little easier to read if you have hierarchical data, but overall the differences are pretty minor in my opinion. And yet - the requests keep rolling in.Hard Link Issues in a Component LibraryAnother reason I've been hesitant is that I really didn't want to pull in a dependency on an external JSON library - in this case JSON.NET - into the core library. If you're not using JSON.NET elsewhere I don't want a user to have to require a hard dependency on JSON.NET unless they want to use the JSON feature. JSON.NET is also sensitive to versions and doesn't play nice with multiple versions when hard linked. For example, when you have a reference to V4.4 in your project but the host application has a reference to version 4.5 you can run into assembly load problems. NuGet's Update-Package can solve some of this *if* you can recompile, but that's not ideal for a component that's supposed to be just plug and play. This is no criticism of JSON.NET - this really applies to any dependency that might change.  So hard linking the DLL can be problematic for a number reasons, but the primary reason is to not force loading of JSON.NET unless you actually need it when you use the JSON configuration features of the library.Enter Dynamic LoadingSo rather than adding an assembly reference to the project, I decided that it would be better to dynamically load the DLL at runtime and then use dynamic typing to access various classes. This allows me to run without a hard assembly reference and allows more flexibility with version number differences now and in the future.But there are also a couple of downsides:No assembly reference means only dynamic access - no compiler type checking or IntellisenseRequirement for the host application to have reference to JSON.NET or else get runtime errorsThe former is minor, but the latter can be problematic. Runtime errors are always painful, but in this case I'm willing to live with this. If you want to use JSON configuration settings JSON.NET needs to be loaded in the project. If this is a Web project, it'll likely be there already.So there are a few things that are needed to make this work:Dynamically create an instance and optionally attempt to load an Assembly (if not loaded)Load types into dynamic variablesUse Reflection for a few tasks like statics/enumsThe dynamic keyword in C# makes the formerly most difficult Reflection part - method calls and property assignments - fairly painless. But as cool as dynamic is it doesn't handle all aspects of Reflection. Specifically it doesn't deal with object activation, truly dynamic (string based) member activation or accessing of non instance members, so there's still a little bit of work left to do with Reflection.Dynamic Object InstantiationThe first step in getting the process rolling is to instantiate the type you need to work with. This might be a two step process - loading the instance from a string value, since we don't have a hard type reference and potentially having to load the assembly. Although the host project might have a reference to JSON.NET, that instance might have not been loaded yet since it hasn't been accessed yet. In ASP.NET this won't be a problem, since ASP.NET preloads all referenced assemblies on AppDomain startup, but in other executable project, assemblies are just in time loaded only when they are accessed.Instantiating a type is a two step process: Finding the type reference and then activating it. Here's the generic code out of my ReflectionUtils library I use for this:/// <summary> /// Creates an instance of a type based on a string. Assumes that the type's /// </summary> /// <param name="typeName">Common name of the type</param> /// <param name="args">Any constructor parameters</param> /// <returns></returns> public static object CreateInstanceFromString(string typeName, params object[] args) { object instance = null; Type type = null; try { type = GetTypeFromName(typeName); if (type == null) return null; instance = Activator.CreateInstance(type, args); } catch { return null; } return instance; } /// <summary> /// Helper routine that looks up a type name and tries to retrieve the /// full type reference in the actively executing assemblies. /// </summary> /// <param name="typeName"></param> /// <returns></returns> public static Type GetTypeFromName(string typeName) { Type type = null; // Let default name binding find it type = Type.GetType(typeName, false); if (type != null) return type; // look through assembly list var assemblies = AppDomain.CurrentDomain.GetAssemblies(); // try to find manually foreach (Assembly asm in assemblies) { type = asm.GetType(typeName, false); if (type != null) break; } return type; } To use this for loading JSON.NET I have a small factory function that instantiates JSON.NET and sets a bunch of configuration settings on the generated object. The startup code also looks for failure and tries loading up the assembly when it fails since that's the main reason the load would fail. Finally it also caches the loaded instance for reuse (according to James the JSON.NET instance is thread safe and quite a bit faster when cached). Here's what the factory function looks like in JsonSerializationUtils:/// <summary> /// Dynamically creates an instance of JSON.NET /// </summary> /// <param name="throwExceptions">If true throws exceptions otherwise returns null</param> /// <returns>Dynamic JsonSerializer instance</returns> public static dynamic CreateJsonNet(bool throwExceptions = true) { if (JsonNet != null) return JsonNet; lock (SyncLock) { if (JsonNet != null) return JsonNet; // Try to create instance dynamic json = ReflectionUtils.CreateInstanceFromString("Newtonsoft.Json.JsonSerializer"); if (json == null) { try { var ass = AppDomain.CurrentDomain.Load("Newtonsoft.Json"); json = ReflectionUtils.CreateInstanceFromString("Newtonsoft.Json.JsonSerializer"); } catch (Exception ex) { if (throwExceptions) throw; return null; } } if (json == null) return null; json.ReferenceLoopHandling = (dynamic) ReflectionUtils.GetStaticProperty("Newtonsoft.Json.ReferenceLoopHandling", "Ignore"); // Enums as strings in JSON dynamic enumConverter = ReflectionUtils.CreateInstanceFromString("Newtonsoft.Json.Converters.StringEnumConverter"); json.Converters.Add(enumConverter); JsonNet = json; } return JsonNet; }This code's purpose is to return a fully configured JsonSerializer instance. As you can see the code tries to create an instance and when it fails tries to load the assembly, and then re-tries loading.Once the instance is loaded some configuration occurs on it. Specifically I set the ReferenceLoopHandling option to not blow up immediately when circular references are encountered. There are a host of other small config setting that might be useful to set, but the default seem to be good enough in recent versions. Note that I'm setting ReferenceLoopHandling which requires an Enum value to be set. There's no real easy way (short of using the cardinal numeric value) to set a property or pass parameters from static values or enums. This means I still need to use Reflection to make this work. I'm using the same ReflectionUtils class I previously used to handle this for me. The function looks up the type and then uses Type.InvokeMember() to read the static property.Another feature I need is have Enum values serialized as strings rather than numeric values which is the default. To do this I can use the StringEnumConverter to convert enums to strings by adding it to the Converters collection.As you can see there's still a bit of Reflection to be done even in C# 4+ with dynamic, but with a few helpers this process is relatively painless.Doing the actual JSON ConversionFinally I need to actually do my JSON conversions. For the Utility class I need serialization that works for both strings and files so I created four methods that handle these tasks two each for serialization and deserialization for string and file.Here's what the File Serialization looks like:/// <summary> /// Serializes an object instance to a JSON file. /// </summary> /// <param name="value">the value to serialize</param> /// <param name="fileName">Full path to the file to write out with JSON.</param> /// <param name="throwExceptions">Determines whether exceptions are thrown or false is returned</param> /// <param name="formatJsonOutput">if true pretty-formats the JSON with line breaks</param> /// <returns>true or false</returns> public static bool SerializeToFile(object value, string fileName, bool throwExceptions = false, bool formatJsonOutput = false) { dynamic writer = null; FileStream fs = null; try { Type type = value.GetType(); var json = CreateJsonNet(throwExceptions); if (json == null) return false; fs = new FileStream(fileName, FileMode.Create); var sw = new StreamWriter(fs, Encoding.UTF8); writer = Activator.CreateInstance(JsonTextWriterType, sw); if (formatJsonOutput) writer.Formatting = (dynamic)Enum.Parse(FormattingType, "Indented"); writer.QuoteChar = '"'; json.Serialize(writer, value); } catch (Exception ex) { Debug.WriteLine("JsonSerializer Serialize error: " + ex.Message); if (throwExceptions) throw; return false; } finally { if (writer != null) writer.Close(); if (fs != null) fs.Close(); } return true; }You can see more of the dynamic invocation in this code. First I grab the dynamic JsonSerializer instance using the CreateJsonNet() method shown earlier which returns a dynamic. I then create a JsonTextWriter and configure a couple of enum settings on it, and then call Serialize() on the serializer instance with the JsonTextWriter that writes the output to disk. Although this code is dynamic it's still fairly short and readable.For full circle operation here's the DeserializeFromFile() version:/// <summary> /// Deserializes an object from file and returns a reference. /// </summary> /// <param name="fileName">name of the file to serialize to</param> /// <param name="objectType">The Type of the object. Use typeof(yourobject class)</param> /// <param name="binarySerialization">determines whether we use Xml or Binary serialization</param> /// <param name="throwExceptions">determines whether failure will throw rather than return null on failure</param> /// <returns>Instance of the deserialized object or null. Must be cast to your object type</returns> public static object DeserializeFromFile(string fileName, Type objectType, bool throwExceptions = false) { dynamic json = CreateJsonNet(throwExceptions); if (json == null) return null; object result = null; dynamic reader = null; FileStream fs = null; try { fs = new FileStream(fileName, FileMode.Open, FileAccess.Read); var sr = new StreamReader(fs, Encoding.UTF8); reader = Activator.CreateInstance(JsonTextReaderType, sr); result = json.Deserialize(reader, objectType); reader.Close(); } catch (Exception ex) { Debug.WriteLine("JsonNetSerialization Deserialization Error: " + ex.Message); if (throwExceptions) throw; return null; } finally { if (reader != null) reader.Close(); if (fs != null) fs.Close(); } return result; }This code is a little more compact since there are no prettifying options to set. Here JsonTextReader is created dynamically and it receives the output from the Deserialize() operation on the serializer.You can take a look at the full JsonSerializationUtils.cs file on GitHub to see the rest of the operations, but the string operations are very similar - the code is fairly repetitive.These generic serialization utilities isolate the dynamic serialization logic that has to deal with the dynamic nature of JSON.NET, and any code that uses these functions is none the wiser that JSON.NET is dynamically loaded.Using the JsonSerializationUtils WrapperThe final consumer of the SerializationUtils wrapper is an actual ConfigurationProvider, that is responsible for handling reading and writing JSON values to and from files. The provider is simple a small wrapper around the SerializationUtils component and there's very little code to make this work now:The whole provider looks like this:/// <summary> /// Reads and Writes configuration settings in .NET config files and /// sections. Allows reading and writing to default or external files /// and specification of the configuration section that settings are /// applied to. /// </summary> public class JsonFileConfigurationProvider<TAppConfiguration> : ConfigurationProviderBase<TAppConfiguration> where TAppConfiguration: AppConfiguration, new() { /// <summary> /// Optional - the Configuration file where configuration settings are /// stored in. If not specified uses the default Configuration Manager /// and its default store. /// </summary> public string JsonConfigurationFile { get { return _JsonConfigurationFile; } set { _JsonConfigurationFile = value; } } private string _JsonConfigurationFile = string.Empty; public override bool Read(AppConfiguration config) { var newConfig = JsonSerializationUtils.DeserializeFromFile(JsonConfigurationFile, typeof(TAppConfiguration)) as TAppConfiguration; if (newConfig == null) { if(Write(config)) return true; return false; } DecryptFields(newConfig); DataUtils.CopyObjectData(newConfig, config, "Provider,ErrorMessage"); return true; } /// <summary> /// Return /// </summary> /// <typeparam name="TAppConfig"></typeparam> /// <returns></returns> public override TAppConfig Read<TAppConfig>() { var result = JsonSerializationUtils.DeserializeFromFile(JsonConfigurationFile, typeof(TAppConfig)) as TAppConfig; if (result != null) DecryptFields(result); return result; } /// <summary> /// Write configuration to XmlConfigurationFile location /// </summary> /// <param name="config"></param> /// <returns></returns> public override bool Write(AppConfiguration config) { EncryptFields(config); bool result = JsonSerializationUtils.SerializeToFile(config, JsonConfigurationFile,false,true); // Have to decrypt again to make sure the properties are readable afterwards DecryptFields(config); return result; } }This incidentally demonstrates how easy it is to create a new provider for the West Wind Application Configuration component. Simply implementing 3 methods will do in most cases.Note this code doesn't have any dynamic dependencies - all that's abstracted away in the JsonSerializationUtils(). From here on, serializing JSON is just a matter of calling the static methods on the SerializationUtils class.Already, there are several other places in some other tools where I use JSON serialization this is coming in very handy. With a couple of lines of code I was able to add JSON.NET support to an older AJAX library that I use replacing quite a bit of code that was previously in use. And for any other manual JSON operations (in a couple of apps I use JSON Serialization for 'blob' like document storage) this is also going to be handy.Performance?Some of you might be thinking that using dynamic and Reflection can't be good for performance. And you'd be right… In performing some informal testing it looks like the performance of the native code is nearly twice as fast as the dynamic code. Most of the slowness is attributable to type lookups. To test I created a native class that uses an actual reference to JSON.NET and performance was consistently around 85-90% faster with the referenced code. That being said though - I serialized 10,000 objects in 80ms vs. 45ms so this isn't hardly slouchy. For the configuration component speed is not that important because both read and write operations typically happen once on first access and then every once in a while. But for other operations - say a serializer trying to handle AJAX requests on a Web Server one would be well served to create a hard dependency.Dynamic Loading - Worth it?On occasion dynamic loading makes sense. But there's a price to be paid in added code complexity and a performance hit. But for some operations that are not pivotal to a component or application and only used under certain circumstances dynamic loading can be beneficial to avoid having to ship extra files and loading down distributions. These days when you create new projects in Visual Studio with 30 assemblies before you even add your own code, trying to keep file counts under control seems a good idea. It's not the kind of thing you do on a regular basis, but when needed it can be a useful tool. Hopefully some of you find this information useful…© Rick Strahl, West Wind Technologies, 2005-2013Posted in .NET  C#   Tweet !function(d,s,id){var js,fjs=d.getElementsByTagName(s)[0];if(!d.getElementById(id)){js=d.createElement(s);js.id=id;js.src="//platform.twitter.com/widgets.js";fjs.parentNode.insertBefore(js,fjs);}}(document,"script","twitter-wjs"); (function() { var po = document.createElement('script'); po.type = 'text/javascript'; po.async = true; po.src = 'https://apis.google.com/js/plusone.js'; var s = document.getElementsByTagName('script')[0]; s.parentNode.insertBefore(po, s); })();

    Read the article

  • How safe is GParted when resizing Linux and Windows partitions?

    - by Olivier Pons
    I want to resize my partitions: I have 3 partitions: Ubuntu 10.04 Windows Seven Ubuntu 11.10 It's booting with the boot installed by the Ubuntu 11.10 version. I want to expand (only expand) all the 3 partitions. My HD is 1,8 Tb so it's big and I have no possibility to save before expanding. So my question is: if you tell me GParted work 99,99 % of the time, I'm willing to take the risk. If you tell me GParted work 90 % of the time, I won't take that risk.

    Read the article

  • Incorrect instructions on Upgrading to 12.10 from 12.04LTS

    - by Russ F
    https://wiki.ubuntu.com/QuantalQuetzal/TechnicalOverview/Beta1 reports incorrect instructions for upgrading from Ubuntu 12.04 LTS. The correct steps are: Alt+F2, Update Manager, choose settings, updates tab and set notify to "For any new version." Close the manager. Press Alt+F2, Terminal, then enter "sudo update-manager -d" (without the quotes)... Sorry to pester this list, but the Ubuntu wiki has no provisions for "Talk" or "Discussion" that do not require registration and a login. I feel like I should be able to point out a problem without signing in.

    Read the article

  • In retrospect, has it been a good idea to use three-valued logic for SQL NULL comparisons?

    - by Heinzi
    In SQL, NULL means "unknown value". Thus, every comparison with NULL yields NULL (unknown) rather than TRUE or FALSE. From a conceptional point of view, this three-valued logic makes sense. From a practical point of view, every learner of SQL has, one time or another, made the classic WHERE myField = NULL mistake or learned the hard way that NOT IN does not do what one would expect when NULL values are present. It is my impression (please correct me if I am wrong) that the cases where this three-valued logic helps (e.g. WHERE myField IS NOT NULL AND myField <> 2 can be shortened to WHERE myField <> 2) are rare and, in those cases, people tend to use the longer version anyway for clarity, just like you would add a comment when using a clever, non-obvious hack. Is there some obvious advantage that I am missing? Or is there a general consensus among the development community that this has been a mistake?

    Read the article

  • DB Documentation Tool

    - by Hisham El-bereky
     Recently I have uploaded new project to codeplex site, DbDocument or DbDoc project is a helper tool used side by side with MS SQL server management studio tool, you can design your DB Tables in visualized way through Diagrams and then use “DbDoc” tool to generate design document in MS Word format, the generated file can be used in design review process or as history reference, the tool facilitate and reduce the time of writing DB structure documentthe current version is not so sophisticated which is intend to generate word document in table format with all tables in DB illustrating its structure and constraints, but for now it seems to be good.   For more details check DbDoc document or go immediately to DbDoc home page http://dbdocument.codeplex.com/

    Read the article

  • Keeping up with LibreOffice

    - by user1768076
    Installing LibreOffice manually on Ubuntu 12.04 works great following the advice in this answer: http://askubuntu.com/a/182602/103893 It basically is a matter of downloading the most recent release from the LibreOffice site and running sudo apt-get remove libreoffice-core and sudo dpkg -i *.deb. Doing the above I installed LibreOffice 3.6.3 and it works great. But what is the best way to uninstall this new version of LibreOffice (for when the next release is coming along)? Trying sudo apt-get remove libreoffice-core will not do anything, and sudo apt-get remove libreoffice3.6 does not seem to remove all packages installed with dpkg...?

    Read the article

  • How do I simulate overprinting in Adobe Reader?

    - by Ben Everard
    On both Mac and Windows when I print a document there is an advanced screen that allows me to select an option called Simulate Overprinting, however such an option doesn't appear on the Ubuntu version. Wikipedia on overprinting: Overprinting refers to the process of printing one colour on top of another in reprographics. This is closely linked to the reprographic technique of 'trapping'. Another use of overprinting is to create a rich black (often regarded as a colour that is "blacker than black") by printing black over another dark colour. This is an issue for us, as we're trying to print documents that need flattening (this is what overprinting does). Am I missing something here, is there a way to enable overprinting on printed PDFs? Note: Please don't confuse simulate overprinting with overprint preview, of which doesn't apply when printing. Just to show you what I'm looking for, this is the Print > Advanced screen... And this is what I see on the Ubuntu screen, not no option for overprinting

    Read the article

  • JPA and NoSQL using EclipseLink - MongoDB supported

    - by arungupta
    EclipseLink 2.4 has added JPA support for NoSQL databases, MongoDB and Oracle NoSQL are the first ones to make the cut. The support to other NoSQL database can be extended by adding a EclipseLink EISPlatform class and a JCA adapter. A Java class can be mapped to a NoSQL datasource using the @NoSQL annotation or <no-sql> XML element. Even a subset of JPQL and the Criteria API are supported, dependent on the NoSQL database's query support. The connection properties are specified in "persistence.xml". A complete sample showing how JPA annotations are mapping and using @NoSQL is explained here. The MongoDB-version of the source code can also be checked out from the SVN repository. EclipseLink 2.4 is scheduled to be released with Eclipse Juno in June 2012 and the complete set of supported features is described on their wiki. The milestone and nightly builds are already available. Do you want to try with GlassFish and let us know ?

    Read the article

  • How can I fix crashing Flash on Chromium?

    - by dv3500ea
    If I try to watch any flash video, I get a notification bubble telling me that the Shockwave Flash plugin has crashed. When run from the terminal, I get this error whenever flash crashes: [xxxx:xxxx:xxxxxxxxxx:FATAL:base/process_util_linux.cc(594)] Out of memory. where the 'x's are seemingly random numbers. Chromium reports its version as 9.0.597.94 (73967). I have tested in other browsers (Midori, Firefox) and flash works properly there. This is a new problem, that I assume was caused by an upgrade. How can I fix this?

    Read the article

  • Recommend good shared hosting [closed]

    - by Django Reinhardt
    It seems that everyone has something bad to say about the "big" shared hosting sites like 1and1, HostGator, GoDaddy, etc. but what are the ones you've had GOOD experiences with? I'm going to focus this question on LAMP stacks, given that they're the most popular option for shared hosting, but if you have an especially good experience with a different stack. Good shared hosting should be: Competitively priced - But not at the expense of... Fully featured - Email, PHP, MySQL, but what else? Highly customizable - Do you have access to advanced features like being able to deliver static content? Up to date - Do they run PHP4 as standard, or do they run the latest version? Customer service - When you have a problem are they rude and unhelpful? Do they take ages to reply? So how about it? Who have YOU have a good experience with?

    Read the article

  • How to mount an ISO (or NRG) image and rip it as if it were a physical CD?

    - by Michael Robinson
    I have an ISO (created from an NRG with nrg2iso) backup of a beloved game from my youth: Dark Reign: Rise of the Shadowhand (1998). I with to relive those better times by listening to game's soundtrack. Is there a way for me to mount said ISO in such a way that I can rip the audio tracks into mp3 files? I ask because although I can successfully mount the ISO, ripit / abcde report no cd inserted. How I mounted the ISO: sudo mount -t iso9660 -o loop iso.iso /media/ISO Alternatively is there another way to recover audio from ISO / NRG images? Update: I was able to mount the NRG version of my backup in a way that ripit recognized using gCDEmu. ripit failed to rip, however, barfing on the first track - which is most definitely a data track. Is there a way to make ripit ignore the first track?

    Read the article

  • What free icons fonts are available? [closed]

    - by Paulocoghi
    With the possibility to embed fonts in websites using @font-face CSS, some creative webdesigners developed their way to display icons using fonts, instead of images. But most of the available icon fonts (or "font-face" kits) are paid, such as: Pictos Fico Tipogram KDN Media IconMoon *free version, with less icons So, what free icons pack in font format do you know? Please limit one "font-face" kit per answer - and include why you enjoy/recommend it. Edit: This question can be converted into a community wiki, if the moderators want, because there is no best answer, and each response is welcome as a knowledge that can be enjoyed by many.

    Read the article

  • Why is my WordPress installation unable to create a directory?

    - by Ankur
    I am a new user of Ubuntu 12.04 64 bit and I want to test WordPress on my local PC installation. I've only installed LAMP and the latest version of WordPress for now. Further actions were creating a directory www in /home, i.e /home/www install wordpress in it. changed Apache to use /home/www instead of /var/www Apache is running smoothly & wordpress seems to be so as well. However, if try to change my background of my wordpress blog it throws an error Unable to create directory wp-content/uploads/2012/12. Is its parent directory writable by the server? Plese help me to resolve my problem.

    Read the article

  • Apple sort la bêta d'iOS 6.1 pour les développeurs, ainsi que la bêta de Xcode 4.6

    Apple sort la bêta d'iOS 6.1 pour les développeurs ainsi que la bêta de Xcode 4.6 À peine avoir propulsé la mise à jour mineure iOS 6.0.1 de son système d'exploitation mobile aux consommateurs, Apple met à la disposition des développeurs la prochaine évolution de sa plateforme. Ceux-ci peuvent dès aujourd'hui télécharger la mise à jour iOS 6.1, afin de découvrir les nouveautés de l'OS et préparer leurs applications avant la sortie grand public de la mouture. iOS 6.1 beta apporte quelques améliorations, surtout à son application native de cartographie Plans, très critiquée dans la version précédente. Il faut noter qu'Apple a mis avant un bouton permettant de soumettre un problème s...

    Read the article

  • DB DOC Enhancements for Oracle SQL Developer v4

    - by thatjeffsmith
    One of our more popular features is ‘DB Doc.’ It’s like JAVADOC for the database. Pick a connection, right-click, and go. It will generate an HTML documentation set for that schema. For version 4, we’ve introduced a few enhancements based on user requests. That’s right, you asked, and we listened. Added support for Package Bodies Added parallelization option for larger doc sets Enhanced the HTML formatting a bit Select Your Object Types and Generation Options We’ve changed the default selection of object types to be included and added support for package bodies There’s also an option to auto-open the documentation set after it’s been generated. And the HTML As Requested

    Read the article

  • Desktop shortcut to create a new desktop shortcut doesn't do anything

    - by David M. Karr
    It's weird that creating desktop shortcuts is currently so primitive. I found the following: Create Shortcut / launcher on Desktop in Ubuntu 12.04 That helps. However, if there's something wrong with the shortcut, it just doesn't do anything. For instance, I tried to create a shortcut for this command line to create a shortcut. When I double-click it, it just does nothing. This is the resulting text of my "Create Desktop Shortcut.desktop" file on the desktop: [Desktop Entry] Version=1.0 Type=Application Terminal=false Icon[en_CA]=gnome-panel-launcher Exec=/usr/bin/gnome-desktop-item-edit ~/Desktop/ --create-new Name[en_CA]=Create Desktop Shortcut Name=Create Desktop Shortcut Icon=gnome-panel-launcher As I said, when I double-click this, or right-click it and select Open, nothing happens. Is there a log file where something about this would be written to?

    Read the article

  • The standards that fail us and the intellectual bubble

    - by Jeff
    There has been a great deal of noise in the techie community about standards, and a sudden and unexplainable hate for Flash. This noise isn't coming from consumers... the countless soccer moms, teens and your weird uncle Bob, it's coming from the people who build (or at least claim to build) the stuff those consumers consume. If you could survey the position of consumers on the topic, they'd likely tell you that they just want stuff on the Web to work.The noise goes something like this: Web standards are the correct and right thing to use across the Intertubes, and anything not a part of those standards (Flash) is bad. Furthermore, the more recent noise is centered around the idea that HTML 5, along with Javascript, is the right thing to use. The arguments against Flash are, well, the truth is I haven't seen a good argument. I see anecdotal nonsense about high CPU usage and things I'd never think to check when I'm watching Piano Cat on YouTube, but these aren't arguments to me. Sure, I've seen it crash a browser a few times, but it's totally rare.But let's go back to standards. Yes, standards have played an important role in establishing the ubiquity of the Web. The protocols themselves, TCP/IP and HTTP, have been critical. HTML, which has served us well for a very long time, established an incredible foundation. Javascript did an OK job, and thanks to clever programmers writing great frameworks like JQuery, is becoming more and more useful. CSS is awful (there, I said it, I feel SO much better), and I'll never understand why it's so disconnected and different from anything else. It doesn't help that it's so widely misinterpreted by different browsers. Still, there's no question that standards are a good thing, and they've been good for the Web, consumers and publishers alike.HTML 4 has been with us for more than a decade. In Web years, that might as well be 80. HTML 5, contrary to popular belief, is not a standard, and likely won't be for many years to come. In fact, the Web hasn't really evolved at all in terms of its standards. The tools that generate the standard markup and script have, but at the end of the day, we're still living with standards that are more than ten years old. The "official" standards process has failed us.The Web evolved anyway, and did not wait for standards bodies to decide what to do next. It evolved in part because Macromedia, then Adobe, kept evolving Flash. In the earlier days, it mostly just did obnoxious splash pages, but then it started doing animation, and then rich apps as they added form input. Eventually it found its killer app: video. Now more than 95% of browsers have Flash installed. Consumers are better for it.But I'll do it one better... I'll go out on a limb and say that Flash is a standard. If it's that pervasive, I don't care what you tell me, it's a standard. Just because a company owns it doesn't mean that it's evil or not a standard. And hey, it pains me to say that as a developer, because I think the dev tools are the suck (more on that in a minute). But again, consumers don't care. They don't even pay for Flash. The bottom line is that if I put something Flash based on the Internet, it's likely that my audience will see it.And what about the speed of standards owned by a company? Look no further than Silverlight. Silverlight 2 (which I consider the "real" start to the story) came out about a year and a half ago. Now version 4 is out, and it has come a very long way in its capabilities. If you believe Riastats.com, more than half of browsers have it now. It didn't have to wait for standards bodies and nerds drafting documents, it's out today. At this rate, Silverlight will be on version 6 or 7 by the time HTML 5 is a ratified standard.Back to the noise, one of the things that has continually disappointed me about this profession is the number of people who get stuck in an intellectual bubble, color it with dogmatic principles, and completely ignore the actual marketplace where this stuff all has to live. We aren't machines; Binary thinking that forces us to choose between "open standards" and "proprietary lock-in" (the most loaded b.s. FUD term evar) isn't smart at all. The truth is that the <object> tag has allowed us to build incredible stuff on top of the old standards, and consumers have benefitted greatly. Consumer desire, capitalism, and yes, standards ratified by nerds who think about this stuff for years have all played a role in the broad adoption of the Interwebs.We could all do without the noise. At the end of the day, I'm going to build stuff for the Web that's good for my users, and I'm not going to base my decisions on a techie bubble religion. Imagine what the brilliant minds behind the noise could do for the Web if they joined me in that pursuit.

    Read the article

< Previous Page | 517 518 519 520 521 522 523 524 525 526 527 528  | Next Page >