Search Results

Search found 5658 results on 227 pages for 'chris 45'.

Page 99/227 | < Previous Page | 95 96 97 98 99 100 101 102 103 104 105 106  | Next Page >

  • Is it possible to create the following XML using Xdocument and append the relevant protion in subseq

    - by Newbie
    <?xml version='1.0' encoding='UTF-8'?> <StockMarket> <StockDate Day = "02" Month="06" Year="2010"> <Stock> <Symbol>ABC</Symbol> <Amount>110.45</Amount> </Stock> <Stock> <Symbol>XYZ</Symbol> <Amount>366.25</Amount> </Stock> </StockDate> <StockDate Day = "03" Month="06" Year="2010"> <Stock> <Symbol>ABC</Symbol> <Amount>110.35</Amount> </Stock> <Stock> <Symbol>XYZ</Symbol> <Amount>369.70</Amount> </Stock> </StockDate> </StockMarket> My approach so far is XDocument doc = new XDocument( new XElement("StockMarket", new XElement("StockDate", new XAttribute("Day", "02"),new XAttribute("Month","06"),new XAttribute("Year","2010")), new XElement("Stock", new XElement("Symbol", "ABC"), new XElement("Amount", "110.45")) ) ); Since I am new to Linq to XML, I am presently struggling a lot and henceforth seeking for help. What I need is to generate the above XML programatically(I mean to say that some kind of looping...) The values will be passed from textbox and hence will be filled at runtime. At present the one which I have done is a kind of static one. This entire stuff has to be done at runtime. One more problem that I am facing is at the time of saving the record for the second time. Suppose first time I executed the code and saved it(Using the program I have done). Next time when I will try to save then only <StockDate Day = "xx" Month="xx" Year="xxxx"> <Stock> <Symbol>ABC</Symbol> <Amount>110.45</Amount> </Stock> <Stock> <Symbol>XYZ</Symbol> <Amount>366.25</Amount> </Stock> </StockDate> should get save(or better appended) and not <StockMarket> .... </StockMarket>.. Because that will be created only at the first time when the XML will be generated or created. I hope I am able to convey my problem properly. If you people is having difficulty in understanding my situation, kindly let me know. Using C#3.0 . Thanks

    Read the article

  • won't repaint a different Month after pressing button in my calendar

    - by DarkStar123
    I'm trying to build a Calendar in Java as a little project I thought of, But I can't seem to change the name of the Month every time I click the Next button. here's my code! package drawing; import javax.swing.*; import java.awt.*; import java.awt.event.*; public class Drawing_something extends JPanel{ int[] calender_squares = {1, 2, 3, 4, 5, 6, 7}; String[] Month = {"January", "February", "March", "April","May","June","July", "August","September","October","November","December"}; int i = 0; Graphics c; @Override public void paintComponent(Graphics c){ super.paintComponent(c); this.setBackground(Color.WHITE); int WIDTH = 55, HEIGHT = 65; for (int in: calender_squares) { for (int counter = 0; counter < 7; counter++){ c.drawRect(50, 50, 100, 100); c.drawRect(50, 50, 700, 500); c.copyArea(50, 50, 600, 500, 100, 0); c.copyArea(50, 50, 600, 400, 0, 100); } } for (int date = 1; date <= 30; date++) { String s = String.valueOf(date); c.drawString(s, WIDTH, HEIGHT); if (date <= 6){ WIDTH += 100; } else if (date == 7){ WIDTH = 55; HEIGHT = 165; }else if (date <= 13){ WIDTH += 100; }else if (date == 14){ WIDTH = 55; HEIGHT = 265; }else if (date <= 20){ WIDTH += 100; }else if (date == 21){ WIDTH = 55; HEIGHT = 365; }else if (date <= 27){ WIDTH += 100; }else if (date == 28){ WIDTH = 55; HEIGHT = 465; }else if (date <= 30){ WIDTH += 100; } } c.setFont(new Font("default", Font.BOLD, 40)); c.drawString(Month[i], 320, 45); } public Drawing_something(){ setLayout(new BorderLayout()); JButton N = new JButton("NEXT"); JButton B = new JButton("BACK"); JPanel P = new JPanel(); P.add(B); P.add(N); add(P, BorderLayout.SOUTH); B.addActionListener(new HandlerClass()); N.addActionListener(new NextClass()); } public class HandlerClass implements ActionListener{ public void actionPerformed(ActionEvent e){ } } public class NextClass implements ActionListener{ public void actionPerformed(ActionEvent e){ if (i == 11){ i = 0; } i = i + 1; c.drawString(Month[i], 320, 45); } } public static void main(String[] args){ JFrame mainFrame = new JFrame("Calender"); mainFrame.add(new Drawing_something()); mainFrame.setSize(850, 650); mainFrame.setDefaultCloseOperation(JFrame.EXIT_ON_CLOSE); mainFrame.setVisible(true); } } if anyone could help that would be much appreciated!! Thanks in advance!!

    Read the article

  • Eclipse editor closing itself

    - by Rohit
    I have been facing this weird problem for some days ... When i try to open my eclipse editor it closes itself and just show some error message that is logged in my .log file .. I am pasting my log file content here .. help ! !ENTRY org.eclipse.jface 4 0 2012-09-23 01:45:17.443 !MESSAGE An error has occurred. See error log for more details. !STACK 0 org.eclipse.core.runtime.AssertionFailedException: assertion failed: The application has not been initialized. at org.eclipse.core.runtime.Assert.isTrue(Assert.java:110) at org.eclipse.core.internal.runtime.InternalPlatform.assertInitialized(InternalPlatform.java:148) at org.eclipse.core.internal.runtime.InternalPlatform.getAdapterManager(InternalPlatform.java:169) at org.eclipse.core.runtime.Platform.getAdapterManager(Platform.java:628) at org.eclipse.ui.part.IntroPart.getAdapter(IntroPart.java:143) at org.eclipse.ui.internal.ViewIntroAdapterPart.getAdapter(ViewIntroAdapterPart.java:117) at org.eclipse.ui.internal.util.Util.getAdapter(Util.java:109) at org.eclipse.ui.internal.services.WorkbenchSourceProvider.getShowInSource(WorkbenchSourceProvider.java:406) at org.eclipse.ui.internal.services.WorkbenchSourceProvider.getContext(WorkbenchSourceProvider.java:410) at org.eclipse.ui.internal.services.WorkbenchSourceProvider.updateActivePart(WorkbenchSourceProvider.java:478) at org.eclipse.ui.internal.services.WorkbenchSourceProvider.checkActivePart(WorkbenchSourceProvider.java:305) at org.eclipse.ui.internal.services.WorkbenchSourceProvider.checkActivePart(WorkbenchSourceProvider.java:300) at org.eclipse.ui.internal.services.WorkbenchSourceProvider$2.windowDeactivated(WorkbenchSourceProvider.java:270) at org.eclipse.ui.internal.Workbench$15.run(Workbench.java:1026) at org.eclipse.core.runtime.SafeRunner.run(SafeRunner.java:42) at org.eclipse.ui.internal.Workbench.fireWindowDeactivated(Workbench.java:1024) at org.eclipse.ui.internal.WorkbenchWindow$29.shellDeactivated(WorkbenchWindow.java:3169) at org.eclipse.swt.widgets.TypedListener.handleEvent(TypedListener.java:111) at org.eclipse.swt.widgets.EventTable.sendEvent(EventTable.java:84) at org.eclipse.swt.widgets.Widget.sendEvent(Widget.java:1258) at org.eclipse.swt.widgets.Widget.sendEvent(Widget.java:1282) at org.eclipse.swt.widgets.Widget.sendEvent(Widget.java:1263) at org.eclipse.swt.widgets.Shell.filterProc(Shell.java:748) at org.eclipse.swt.widgets.Display.filterProc(Display.java:1543)

    Read the article

  • ASP.NET MVC ‘Extendable-hooks’ – ControllerActionInvoker class

    - by nmarun
    There’s a class ControllerActionInvoker in ASP.NET MVC. This can be used as one of an hook-points to allow customization of your application. Watching Brad Wilsons’ Advanced MP3 from MVC Conf inspired me to write about this class. What MSDN says: “Represents a class that is responsible for invoking the action methods of a controller.” Well if MSDN says it, I think I can instill a fair amount of confidence into what the class does. But just to get to the details, I also looked into the source code for MVC. Seems like the base class Controller is where an IActionInvoker is initialized: 1: protected virtual IActionInvoker CreateActionInvoker() { 2: return new ControllerActionInvoker(); 3: } In the ControllerActionInvoker (the O-O-B behavior), there are different ‘versions’ of InvokeActionMethod() method that actually call the action method in question and return an instance of type ActionResult. 1: protected virtual ActionResult InvokeActionMethod(ControllerContext controllerContext, ActionDescriptor actionDescriptor, IDictionary<string, object> parameters) { 2: object returnValue = actionDescriptor.Execute(controllerContext, parameters); 3: ActionResult result = CreateActionResult(controllerContext, actionDescriptor, returnValue); 4: return result; 5: } I guess that’s enough on the ‘behind-the-screens’ of this class. Let’s see how we can use this class to hook-up extensions. Say I have a requirement that the user should be able to get different renderings of the same output, like html, xml, json, csv and so on. The user will type-in the output format in the url and should the get result accordingly. For example: http://site.com/RenderAs/ – renders the default way (the razor view) http://site.com/RenderAs/xml http://site.com/RenderAs/csv … and so on where RenderAs is my controller. There are many ways of doing this and I’m using a custom ControllerActionInvoker class (even though this might not be the best way to accomplish this). For this, my one and only route in the Global.asax.cs is: 1: routes.MapRoute("RenderAsRoute", "RenderAs/{outputType}", 2: new {controller = "RenderAs", action = "Index", outputType = ""}); Here the controller name is ‘RenderAsController’ and the action that’ll get called (always) is the Index action. The outputType parameter will map to the type of output requested by the user (xml, csv…). I intend to display a list of food items for this example. 1: public class Item 2: { 3: public int Id { get; set; } 4: public string Name { get; set; } 5: public Cuisine Cuisine { get; set; } 6: } 7:  8: public class Cuisine 9: { 10: public int CuisineId { get; set; } 11: public string Name { get; set; } 12: } Coming to my ‘RenderAsController’ class. I generate an IList<Item> to represent my model. 1: private static IList<Item> GetItems() 2: { 3: Cuisine cuisine = new Cuisine { CuisineId = 1, Name = "Italian" }; 4: Item item = new Item { Id = 1, Name = "Lasagna", Cuisine = cuisine }; 5: IList<Item> items = new List<Item> { item }; 6: item = new Item {Id = 2, Name = "Pasta", Cuisine = cuisine}; 7: items.Add(item); 8: //... 9: return items; 10: } My action method looks like 1: public IList<Item> Index(string outputType) 2: { 3: return GetItems(); 4: } There are two things that stand out in this action method. The first and the most obvious one being that the return type is not of type ActionResult (or one of its derivatives). Instead I’m passing the type of the model itself (IList<Item> in this case). We’ll convert this to some type of an ActionResult in our custom controller action invoker class later. The second thing (a little subtle) is that I’m not doing anything with the outputType value that is passed on to this action method. This value will be in the RouteData dictionary and we’ll use this in our custom invoker class as well. It’s time to hook up our invoker class. First, I’ll override the Initialize() method of my RenderAsController class. 1: protected override void Initialize(RequestContext requestContext) 2: { 3: base.Initialize(requestContext); 4: string outputType = string.Empty; 5:  6: // read the outputType from the RouteData dictionary 7: if (requestContext.RouteData.Values["outputType"] != null) 8: { 9: outputType = requestContext.RouteData.Values["outputType"].ToString(); 10: } 11:  12: // my custom invoker class 13: ActionInvoker = new ContentRendererActionInvoker(outputType); 14: } Coming to the main part of the discussion – the ContentRendererActionInvoker class: 1: public class ContentRendererActionInvoker : ControllerActionInvoker 2: { 3: private readonly string _outputType; 4:  5: public ContentRendererActionInvoker(string outputType) 6: { 7: _outputType = outputType.ToLower(); 8: } 9: //... 10: } So the outputType value that was read from the RouteData, which was passed in from the url, is being set here in  a private field. Moving to the crux of this article, I now override the CreateActionResult method. 1: protected override ActionResult CreateActionResult(ControllerContext controllerContext, ActionDescriptor actionDescriptor, object actionReturnValue) 2: { 3: if (actionReturnValue == null) 4: return new EmptyResult(); 5:  6: ActionResult result = actionReturnValue as ActionResult; 7: if (result != null) 8: return result; 9:  10: // This is where the magic happens 11: // Depending on the value in the _outputType field, 12: // return an appropriate ActionResult 13: switch (_outputType) 14: { 15: case "json": 16: { 17: JavaScriptSerializer serializer = new JavaScriptSerializer(); 18: string json = serializer.Serialize(actionReturnValue); 19: return new ContentResult { Content = json, ContentType = "application/json" }; 20: } 21: case "xml": 22: { 23: XmlSerializer serializer = new XmlSerializer(actionReturnValue.GetType()); 24: using (StringWriter writer = new StringWriter()) 25: { 26: serializer.Serialize(writer, actionReturnValue); 27: return new ContentResult { Content = writer.ToString(), ContentType = "text/xml" }; 28: } 29: } 30: case "csv": 31: controllerContext.HttpContext.Response.AddHeader("Content-Disposition", "attachment; filename=items.csv"); 32: return new ContentResult 33: { 34: Content = ToCsv(actionReturnValue as IList<Item>), 35: ContentType = "application/ms-excel" 36: }; 37: case "pdf": 38: string filePath = controllerContext.HttpContext.Server.MapPath("~/items.pdf"); 39: controllerContext.HttpContext.Response.AddHeader("content-disposition", 40: "attachment; filename=items.pdf"); 41: ToPdf(actionReturnValue as IList<Item>, filePath); 42: return new FileContentResult(StreamFile(filePath), "application/pdf"); 43:  44: default: 45: controllerContext.Controller.ViewData.Model = actionReturnValue; 46: return new ViewResult 47: { 48: TempData = controllerContext.Controller.TempData, 49: ViewData = controllerContext.Controller.ViewData 50: }; 51: } 52: } A big method there! The hook I was talking about kinda above actually is here. This is where different kinds / formats of output get returned based on the output type requested in the url. When the _outputType is not set (string.Empty as set in the Global.asax.cs file), the razor view gets rendered (lines 45-50). This is the default behavior in most MVC applications where-in a view (webform/razor) gets rendered on the browser. As you see here, this gets returned as a ViewResult. But then, for an outputType of json/xml/csv, a ContentResult gets returned, while for pdf, a FileContentResult is returned. Here are how the different kinds of output look like: This is how we can leverage this feature of ASP.NET MVC to developer a better application. I’ve used the iTextSharp library to convert to a pdf format. Mike gives quite a bit of detail regarding this library here. You can download the sample code here. (You’ll get an option to download once you open the link). Verdict: Hot chocolate: $3; Reebok shoes: $50; Your first car: $3000; Being able to extend a web application: Priceless.

    Read the article

  • Event on SQL Server 2008 Disk IO and the new Complex Event Processing (StreamInsight) feature in R2

    - by tonyrogerson
    Allan Mitchell and myself are doing a double act, Allan is becoming one of the leading guys in the UK on StreamInsight and will give an introduction to this new exciting technology; on top of that I'll being talking about SQL Server Disk IO - well, "Disk" might not be relevant anymore because I'll be talking about SSD and IOFusion - basically I'll be talking about the underpinnings - making sure you understand and get it right, how to monitor etc... If you've any specific problems or questions just ping me an email [email protected]. To register for the event see: http://sqlserverfaq.com/events/217/SQL-Server-and-Disk-IO-File-GroupsFiles-SSDs-FusionIO-InRAM-DBs-Fragmentation-Tony-Rogerson-Complex-Event-Processing-Allan-Mitchell.aspx 18:15 SQL Server and Disk IOTony Rogerson, SQL Server MVPTony's Blog; Tony on TwitterIn this session Tony will talk about RAID levels, how SQL server writes to and reads from disk, the effect SSD has and will talk about other options for throughput enhancement like Fusion IO. He will look at the effect fragmentation has and how to minimise the impact, he will look at the File structure of a database and talk about what benefits multiple files and file groups bring. We will also touch on Database Mirroring and the effect that has on throughput, how to get a feeling for the throughput you should expect.19:15 Break19:45 Complex Event Processing (CEP)Allan Mitchell, SQL Server MVPhttp://sqlis.com/sqlisStreamInsight is Microsoft’s first foray into the world of Complex Event Processing (CEP) and Event Stream Processing (ESP).  In this session I want to show an introduction to this technology.  I will show how and why it is useful.  I will get us used to some new terminology but best of all I will show just how easy it is to start building your first CEP/ESP application.

    Read the article

  • SQL SERVER – What is Page Life Expectancy (PLE) Counter

    - by pinaldave
    During performance tuning consultation there are plenty of counters and values, I often come across. Today we will quickly talk about Page Life Expectancy counter, which is commonly known as PLE as well. You can find the value of the PLE by running following query. SELECT [object_name], [counter_name], [cntr_value] FROM sys.dm_os_performance_counters WHERE [object_name] LIKE '%Manager%' AND [counter_name] = 'Page life expectancy' The recommended value of the PLE counter is 300 seconds. I have seen on busy system this value to be as low as even 45 seconds and on unused system as high as 1250 seconds. Page Life Expectancy is number of seconds a page will stay in the buffer pool without references. In simple words, if your page stays longer in the buffer pool (area of the memory cache) your PLE is higher, leading to higher performance as every time request comes there are chances it may find its data in the cache itself instead of going to hard drive to read the data. Now check your system and post back what is this counter value for you during various time of the day. Is this counter any way relates to performance issues for your system? Note: There are various other counters which are important to discuss during the performance tuning and this counter is not everything. Reference: Pinal Dave (http://blog.SQLAuthority.com) Filed under: Pinal Dave, SQL, SQL Authority, SQL Optimization, SQL Performance, SQL Query, SQL Scripts, SQL Server, SQL Tips and Tricks, SQLServer, T SQL, Technology

    Read the article

  • Confusion for mime files: magic, magic.mgc, magic.mime

    - by Florence Foo
    I'm using Ubuntu. I'm trying to use ruby gem 'shared-mime-info' for an application I'm writing. I understand that magic.mgc is a compiled version of magic file which has magic number definitions for the different file types. BUT I don't understand why is it /usr/share/mime/magic is in binary format instead of just normal text file with each parameters separated by white space like everywhere else I'm finding on the internet when it's referencing this file? The /usr/share/mime/magic has the word 'MIME-Magic' at the beginning of the file and prioritize the rest of the stuff like. So it doesn't look like magic.mgc at all. [100:application/vnd.scribus] >1=^@^KSCRIBUSUTF8 [90:application/vnd.stardivision.writer] >2089=^@ shared-mime-info seems to want a magic file in the binary non compiled format as above and I wanted to add definition for DOCX but how does one update or generate this file without using a hex editor? There is a reference to the magic file I found at: http://standards.freedesktop.org/shared-mime-info-spec/shared-mime-info-spec-latest.html And it mention this file is updated with update-mime-database but what if I just want to add some new entry to it. hex editor? Anyway I ended up using hexer to make a new magic file in ~/.local/share/mime/ with only the entry I wanted to add and the MIME-Magic header. Seems to work (assuming I will ever deal with docx for now). 00000000: 4d 49 4d 45 2d 4d 61 67 69 63 00 0a 5b 36 30 3a MIME-Magic..[60: 00000010: 61 70 70 6c 69 63 61 74 69 6f 6e 2f 76 6e 64 2e application/vnd. 00000020: 6f 70 65 6e 78 6d 6c 66 6f 72 6d 61 74 73 2d 6f openxmlformats-o 00000030: 66 66 69 63 65 64 6f 63 75 6d 65 6e 74 2e 77 6f fficedocument.wo 00000040: 72 64 70 72 6f 63 65 73 73 69 6e 67 6d 6c 2e 64 rdprocessingml.d 00000050: 6f 63 75 6d 65 6e 74 5d 0a 3e 30 3d 00 08 50 4b ocument].>0=..PK 00000060: 03 04 14 00 06 00 0a -- -- -- -- -- -- -- -- -- .......---------

    Read the article

  • Join us for our WebLogic Communtiy webcast on November 2nd 2012! OOW update WebLogic & ExaLogic

    - by JuergenKress
    NOVEMBER 2nd, 2012 AT 11:00 -11:45 AM CET (BERLIN TIME) Do you want to learn the latest about WebLogic and ExaLogic? Would you like to know what did Oracle announce at Oracle OpenWorld 2012? Join us for our WebLogic Communtiy webcast on November 2nd 2012! Don't miss this unique opportunity and learn the latest announcement and product updates on WebLogic and ExaLogic portfolio. Agenda Update announced about Cloud Application Foundation ExaLogic Update Key examples from successful customer Updated Roadmap Q&A Register to Attend! To Join the Webcast (Employees and Partners) AND / OR DIAL IN If you would like to dial in to the audio portion of the Webconference. Call ID: 8000524 & Call Passcode: 333111 Austria : +43(0)19286512 Belgium: +32(0)24010528 Denmark: +4532729222 Finland: +358(0)923193923 France: +33(0)176728936 Germany: +49(0)69222216106 Ireland: +353(0)12475650 Italy: +39(0)236008198 Netherlands: +31(0)207143543 Norway: +4721033443 Spain: +34914143755 Sweden: +46(0)856619465 Switzerland: +41(0)445804003 United Kingdom: +44(0)2081181001 Please click here to find more Local Numbers. Presenters: Maciej Gruszka Senior Principal Product Manager Tel: +4 (0) 8601 156 464 E-Mail: [email protected] LinkedIn: http://pl.linkedin.com/pub/maciej-gruszka/2/169/89 Jürgen Kress WebLogic Partner Adoption EMEA Tel. +49 89 1430 1479 E-Mail: [email protected] LinkedIn: http://de.linkedin.com/in/kress Twitter: http://twitter.com/wlscommunity Delivery Format This FREE online LIVE eSeminar will be delivered over the Web. Registrations received less than 24hours prior to start time may not receive confirmation to attend. Duration: 1 hour Looking forward to your participation! WebLogic Partner Community For regular information become a member in the WebLogic Partner Community please visit: http://www.oracle.com/partners/goto/wls-emea ( OPN account required). If you need support with your account please contact the Oracle Partner Business Center. BlogTwitterLinkedInMixForumWiki Technorati Tags: WebLogic Community Webcast,webcst,OOW updates WebLogic ExaLogic,WebLogic ppt,presenation,sales,WebLogic Community,Oracle,OPN,Jürgen Kress

    Read the article

  • Post build events using ROBOCOPY instead of XCOPY

    - by Vizioz Limited
    I don't know about you, but for a long time I have used XCOPY statements in my Visual Studio post build events to copy my Umbraco files from the project folders to the local version of the website associated with the project.For the last few months we have been building a website framework for a client, who has subsequently sold the site to 5 clients, each with a different skin and some variations in their functional requirements.So, we now have a single source solutions, that builds and copies the site files into 5 seperate local websites, which enables us to easily test them all, what we had found was that this process was starting to slow up our build process and was reaching 30-45 seconds on a high spec Quad core machine (and slower on others)Today I asked Colin to create seperate Solution Configurations within Visual Studio so that while we were developing we could target a single site, and when we wanted to test all sites, we could target "ALL" and the Post Build script would then copy the files to all sites.This worked well, and with a couple of other optimisations, our build was now taking about 10 seconds for a single site.Then Colin came across ROBOCOPY and suggested that maybe this would be a suitable alternative to XCOPY, well, I had not heard of it.. (shock horror some of you shout, some I am sure like me, are also wondering what it is!)ROBOCOPY is new in Windows Vista & Windows 7 (you can also download it for XP & Windows 2003) and it has a lot of additional features, the two that were most interesting to us were:/MIR = Mirror a folder tree/XD = Exclude Directories/NP = No Progress (i.e. it does not give you a chart of it's results, which just fills up your Output window!)So, we set about implementing ROBOCOPY, we decided to use the /MIR switch on all folders that we knew were always stored in our project folders:- images- css- masterpages- xsltAnd for other files we just used the straight robocopy functionality.We also decided to exclude all the .SVN directories using the /XD switch and finally we added the /NP switch as mentioned above.The beauty of all of this, is the /MIR functionality, as this means that only files that have changed will be copied across which greatly speeds up the process, especially on the images folders which previously copied across on every build, now, if we add a new image to the project it will be copied across automatically and then never again, unless we change it of course!The build time now for all sites is approximately 4 seconds and for a single site, 2 seconds, I would highly recommend the time to make the same optimisations to your build processes if you have not done so already.

    Read the article

  • Disable Ethernet permanently to speed up boot time

    - by Anwar Shah
    I do not use the wired Ethernet Card. It seems to me that, Ubuntu is always trying in boot time to check the network via eth0, Which consumes some times and I guess this may slow down the boot process a bit. My dmesg output is below (partial) 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 1.985592] input: Video Bus as /devices/LNXSYSTM:00/device:00/PNP0A08:00/LNXVIDEO:01/input/input5 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 1.985651] ACPI: Video Device [GFX0] (multi-head: yes rom: no post: no) 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 1.985693] [drm] Initialized i915 1.6.0 20080730 for 0000:00:02.0 on minor 0 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 2.056261] firewire_core: created device fw0: GUID 00023f87af41fd7d, S400 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 3.710435] EXT4-fs (sda9): mounted filesystem with ordered data mode. Opts: (null) A big time here..... 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 13.466642] ADDRCONF(NETDEV_UP): eth0: link is not ready 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 14.125296] Adding 1050620k swap on /dev/sda6. Priority:-1 extents:1 across:1050620k 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 14.226952] EXT4-fs (sda9): re-mounted. Opts: (null) 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 14.335012] snd_hda_intel 0000:00:1b.0: PCI INT A - GSI 22 (level, low) - IRQ 22 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 14.335091] snd_hda_intel 0000:00:1b.0: irq 45 for MSI/MSI-X 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 14.335128] snd_hda_intel 0000:00:1b.0: setting latency timer to 64 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 14.346410] input: Ideapad extra buttons as /devices/platform/ideapad/input/input6 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 14.428551] input: HDA Intel Headphone as /devices/pci0000:00/0000:00:1b.0/sound/card0/input7 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 14.436958] cfg80211: Calling CRDA to update world regulatory domain 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 14.476550] Linux video capture interface: v2.00 2012-06-11 23:06:47 Ubuntu-KDE kernel [ 14.486385] uvcvideo: Found UVC 1.00 device USB 2.0 Camera (04f2:b008) So, My question is How can I disable the Ethernet card completely, so that kernel will not try to use that?

    Read the article

  • Why are marketing employees, product managers, etc. deserving of their own office, yet programmers are jammed in a room as many as possible?

    - by TheImirOfGroofunkistan
    I don't understand why many (many) companies treat software developers like they are assembly line workers making widgets. Joel Spolsky has a great example of the problems this creates: With programmers, it's especially hard. Productivity depends on being able to juggle a lot of little details in short term memory all at once. Any kind of interruption can cause these details to come crashing down. When you resume work, you can't remember any of the details (like local variable names you were using, or where you were up to in implementing that search algorithm) and you have to keep looking these things up, which slows you down a lot until you get back up to speed. Here's the simple algebra. Let's say (as the evidence seems to suggest) that if we interrupt a programmer, even for a minute, we're really blowing away 15 minutes of productivity. For this example, lets put two programmers, Jeff and Mutt, in open cubicles next to each other in a standard Dilbert veal-fattening farm. Mutt can't remember the name of the Unicode version of the strcpy function. He could look it up, which takes 30 seconds, or he could ask Jeff, which takes 15 seconds. Since he's sitting right next to Jeff, he asks Jeff. Jeff gets distracted and loses 15 minutes of productivity (to save Mutt 15 seconds). Now let's move them into separate offices with walls and doors. Now when Mutt can't remember the name of that function, he could look it up, which still takes 30 seconds, or he could ask Jeff, which now takes 45 seconds and involves standing up (not an easy task given the average physical fitness of programmers!). So he looks it up. So now Mutt loses 30 seconds of productivity, but we save 15 minutes for Jeff. Ahhh! Quote Link More Spolsky on Offices Why don't managers and owner's see this?

    Read the article

  • Why do marketing employees get their own office, yet programmers are jammed in a room as many as possible?

    - by TheImirOfGroofunkistan
    I don't understand why many (many) companies treat software developers like they are assembly line workers making widgets. Joel Spolsky has a great example of the problems this creates: With programmers, it's especially hard. Productivity depends on being able to juggle a lot of little details in short term memory all at once. Any kind of interruption can cause these details to come crashing down. When you resume work, you can't remember any of the details (like local variable names you were using, or where you were up to in implementing that search algorithm) and you have to keep looking these things up, which slows you down a lot until you get back up to speed. Here's the simple algebra. Let's say (as the evidence seems to suggest) that if we interrupt a programmer, even for a minute, we're really blowing away 15 minutes of productivity. For this example, lets put two programmers, Jeff and Mutt, in open cubicles next to each other in a standard Dilbert veal-fattening farm. Mutt can't remember the name of the Unicode version of the strcpy function. He could look it up, which takes 30 seconds, or he could ask Jeff, which takes 15 seconds. Since he's sitting right next to Jeff, he asks Jeff. Jeff gets distracted and loses 15 minutes of productivity (to save Mutt 15 seconds). Now let's move them into separate offices with walls and doors. Now when Mutt can't remember the name of that function, he could look it up, which still takes 30 seconds, or he could ask Jeff, which now takes 45 seconds and involves standing up (not an easy task given the average physical fitness of programmers!). So he looks it up. So now Mutt loses 30 seconds of productivity, but we save 15 minutes for Jeff. Ahhh! Quote Link More Spolsky on Offices Why don't managers and owner's see this?

    Read the article

  • UBJsonReader (Libgdx) unable to to read UBJson from Python(Blender)

    - by daniel
    I am working on an export tool from Blender to Libgdx, exports like custom attributes and other information (Almost completed), this is a very cool tool that will speed up a lot your works, after I completed I will send to public to contribute forum, Export format is uses python's Standard Json module and readable text, it of course works fine, but I wanna also have a Binary Json export for faster load, so users can Export Straight to Libgdx, but after I search I found that UBJson with draft9.py (simpleubjson 0.6.1) encode is seems matches with one FBXConverter's UBJsonWriter( Xoppa wrote), but when I export, I am not able to read the file, and send this errors (Java heap space) seems this is a different between byte sizes in UBJson(python) and UBJsonReader. how can I write a correct one in python that matches with Libgdx's UBJsonReader, and would be cross-platform? Exception in thread "LWJGL Application" com.badlogic.gdx.utils.GdxRuntimeException: java.lang.OutOfMemoryError: Java heap space at com.badlogic.gdx.backends.lwjgl.LwjglApplication$1.run(LwjglApplication.java:120) Caused by: java.lang.OutOfMemoryError: Java heap space at com.badlogic.gdx.utils.UBJsonReader.readString(UBJsonReader.java:162) at com.badlogic.gdx.utils.UBJsonReader.parseString(UBJsonReader.java:150) at com.badlogic.gdx.utils.UBJsonReader.parseObject(UBJsonReader.java:112) at com.badlogic.gdx.utils.UBJsonReader.parse(UBJsonReader.java:59) at com.badlogic.gdx.utils.UBJsonReader.parse(UBJsonReader.java:52) at com.badlogic.gdx.utils.UBJsonReader.parse(UBJsonReader.java:36) at com.badlogic.gdx.utils.UBJsonReader.parse(UBJsonReader.java:45) at com.me.gdximportexport.GdxImportExport.create(GdxImportExport.java:43) at com.badlogic.gdx.backends.lwjgl.LwjglApplication.mainLoop(LwjglApplication.java:136) at com.badlogic.gdx.backends.lwjgl.LwjglApplication$1.run(LwjglApplication.java:114) Tested on UbuntuStudio 13.10 with OpenJdk 7, and Windows 7 with jdk 7 Thanks for any guides.

    Read the article

  • Watch IPL League 2010 Online On YouTube

    - by Suganya
    Having said that IPL League match for the year 2010 starts tomorrow, Many of us would be interested in watching the broadcast live online. The first match between Deccan chargers and Kolkata Knight Riders starts tomorrow at 20.00 IST. IPL T20 lasts for 45 days starting from March 12th 2010 till April 25th 2010. The entire league takes place in India. The opening ceremony takes place on 12th continued by the official game between Deccan Chargers and Kolkata Knight Riders Most of us would not be able to watch the match on Television. So this year IPL joins hand with Google to make it available live on NET. How to Watch IPL Match Live Online Google and IPL has joined their hands to make the match available online for all the viewers around the world. To watch the IPL Match live online log on here How to Watch IPL Match Live On Mobile IPL and GCV(Global Cricket Ventures) are tied to July Systems. You can watch IPL matches live on mobile by accessing the link directly using a GPRS enabled mobile or viewers can simply call from their mobile phone to MiX (Mobile Internet Experience) using the Toll Free number 08 123 123 123. Once you call this Toll Free number a link will be sent to your mobile. You can access that link to view the live cricket on your mobile. Watch IPL League Cricket Match Live Online Watch IPL League Cricket Match Live On Mobile Join us on Facebook to read all our stories right inside your Facebook news feed.

    Read the article

  • The Social Business Thought Leaders - John Hagel

    - by kellsey.ruppel
    While many European economies are on the brink of a recession between increasing taxation and mounting loss of jobs and bankruptcy filing rates, there's an understandable risk of losing sight of the deeper forces at play. Yet instead of surrendering to uncertainty and trying to survive in the short term, many organizations are feeling the urge to be better prepared to thrive in these complex times by developing a more articulated long term understanding of both the opportunities / challenges ahead. For example: What long-term economic, technological and societal changes are rolling out? Which foundational dynamics will affect our companies' performance, productivity, competition, and innovative potential in the upcoming decades? How will digital infrastructure change our business landscape? What kind of capabilities will be key to compete in a market shaped by growing turbulence, unpredictability and volatility? Breaking out from a strictly cyclical thinking, studies such as the Shift Index by John Hagel, Co-Chairman of the Center for the Edge at Deloitte & Touche (See Measuring the forces of long-term change - The 2009 Shift Index), depict a worrying performance challenge that affected every industry in the entire US economy over the last 45 years. Amidst a more than doubled competitive intensity of the market, and even with an improved labor productivity, the actual performance of US firms has consistently fallen to 25% of what it was in 1965. Most of this reported value is shifting from institutions and organizations to individuals, whether they are customers or young creative talent. To thrive in the digital economy and reverse declining performance trends, companies will have to fundamentally rethink their management approach by moving from knowledge stocks to knowledge flows, from scalable efficiency to scalable learning, from push organizations to pull organizations. Based on the outcomes of the Shift Index and on the book The Power of Pull, the first episode of the Social Business Thought-Leaders features John Hagel to provide strategic insights on how companies will succeed in the 21st century.

    Read the article

  • Free Certification Exams for Visual Studio 2010

    - by budugu
    Get promotional codes from herehttp://blogs.msdn.com/gerryo/archive/2010/03/17/register-for-visual-studio-2010-beta-exams.aspx You don’t have to pay anything to take these exams.  These are 100% free. If you pass the exam, you earn the certification just the same as if you took it in a non-beta environment. From Gerry O'Brien’s blog...  2) Is this a real exam? – Yes it is.  Even though the questions are not scored at the time you take the exam, they are real questions and the exam is real.  If you pass the exam, you earn the certification just the same as if you took it in a non-beta environment.  This means you don’t get a pass/fail or score immediately following the exam, but you do get notified 8 to 10 weeks later because we move slow in getting the final scoring in place.  4) What is the main difference between a beta and non-beta exam, besides cost? – The beta exam will show you questions that have not been through a final QA check.  You are that final QA check.  Non-beta exams expose you to 40 or 45 questions and you have a total of two hours to complete it.  The beta exam could expose you to as many as 125 to 150 questions and take up to four hours.   Following exams are for Asp.Net developers Exam 71-515, TS: Web Applications Development with Microsoft .NET Framework 4Exam 71-519: Pro: Designing and Developing Web Applications Using Microsoft .NET Framework 4

    Read the article

  • NHibernate Conventions

    - by Ricardo Peres
    Introduction It seems that nowadays everyone loves conventions! Not the ones that you go to, but the ones that you use, that is! It just happens that NHibernate also supports conventions, and we’ll see exactly how. Conventions in NHibernate are supported in two ways: Naming of tables and columns when not explicitly indicated in the mappings; Full domain mapping. Naming of Tables and Columns Since always NHibernate has supported the concept of a naming strategy. A naming strategy in NHibernate converts class and property names to table and column names and vice-versa, when a name is not explicitly supplied. In concrete, it must be a realization of the NHibernate.Cfg.INamingStrategy interface, of which NHibernate includes two implementations: DefaultNamingStrategy: the default implementation, where each column and table are mapped to identically named properties and classes, for example, “MyEntity” will translate to “MyEntity”; ImprovedNamingStrategy: underscores (_) are used to separate Pascal-cased fragments, for example, entity “MyEntity” will be mapped to a “my_entity” table. The naming strategy can be defined at configuration level (the Configuration instance) by calling the SetNamingStrategy method: 1: cfg.SetNamingStrategy(ImprovedNamingStrategy.Instance); Both the DefaultNamingStrategy and the ImprovedNamingStrategy classes offer singleton instances in the form of Instance static fields. DefaultNamingStrategy is the one NHibernate uses, if you don’t specify one. Domain Mapping In mapping by code, we have the choice of relying on conventions to do the mapping automatically. This means a class will inspect our classes and decide how they will relate to the database objects. The class that handles conventions is NHibernate.Mapping.ByCode.ConventionModelMapper, a specialization of the base by code mapper, NHibernate.Mapping.ByCode.ModelMapper. The ModelMapper relies on an internal SimpleModelInspector to help it decide what and how to map, but the mapper lets you override its decisions.  You apply code conventions like this: 1: //pick the types that you want to map 2: IEnumerable<Type> types = Assembly.GetExecutingAssembly().GetExportedTypes(); 3:  4: //conventions based mapper 5: ConventionModelMapper mapper = new ConventionModelMapper(); 6:  7: HbmMapping mapping = mapper.CompileMappingFor(types); 8:  9: //the one and only configuration instance 10: Configuration cfg = ...; 11: cfg.AddMapping(mapping); This is a very simple example, it lacks, at least, the id generation strategy, which you can add by adding an event handler like this: 1: mapper.BeforeMapClass += (IModelInspector modelInspector, Type type, IClassAttributesMapper classCustomizer) => 2: { 3: classCustomizer.Id(x => 4: { 5: //set the hilo generator 6: x.Generator(Generators.HighLow); 7: }); 8: }; The mapper will fire events like this whenever it needs to get information about what to do. And basically this is all it takes to automatically map your domain! It will correctly configure many-to-one and one-to-many relations, choosing bags or sets depending on your collections, will get the table and column names from the naming strategy we saw earlier and will apply the usual defaults to all properties, such as laziness and fetch mode. However, there is at least one thing missing: many-to-many relations. The conventional mapper doesn’t know how to find and configure them, which is a pity, but, alas, not difficult to overcome. To start, for my projects, I have this rule: each entity exposes a public property of type ISet<T> where T is, of course, the type of the other endpoint entity. Extensible as it is, NHibernate lets me implement this very easily: 1: mapper.IsOneToMany((MemberInfo member, Boolean isLikely) => 2: { 3: Type sourceType = member.DeclaringType; 4: Type destinationType = member.GetMemberFromDeclaringType().GetPropertyOrFieldType(); 5:  6: //check if the property is of a generic collection type 7: if ((destinationType.IsGenericCollection() == true) && (destinationType.GetGenericArguments().Length == 1)) 8: { 9: Type destinationEntityType = destinationType.GetGenericArguments().Single(); 10:  11: //check if the type of the generic collection property is an entity 12: if (mapper.ModelInspector.IsEntity(destinationEntityType) == true) 13: { 14: //check if there is an equivalent property on the target type that is also a generic collection and points to this entity 15: PropertyInfo collectionInDestinationType = destinationEntityType.GetProperties().Where(x => (x.PropertyType.IsGenericCollection() == true) && (x.PropertyType.GetGenericArguments().Length == 1) && (x.PropertyType.GetGenericArguments().Single() == sourceType)).SingleOrDefault(); 16:  17: if (collectionInDestinationType != null) 18: { 19: return (false); 20: } 21: } 22: } 23:  24: return (true); 25: }); 26:  27: mapper.IsManyToMany((MemberInfo member, Boolean isLikely) => 28: { 29: //a relation is many to many if it isn't one to many 30: Boolean isOneToMany = mapper.ModelInspector.IsOneToMany(member); 31: return (!isOneToMany); 32: }); 33:  34: mapper.BeforeMapManyToMany += (IModelInspector modelInspector, PropertyPath member, IManyToManyMapper collectionRelationManyToManyCustomizer) => 35: { 36: Type destinationEntityType = member.LocalMember.GetPropertyOrFieldType().GetGenericArguments().First(); 37: //set the mapping table column names from each source entity name plus the _Id sufix 38: collectionRelationManyToManyCustomizer.Column(destinationEntityType.Name + "_Id"); 39: }; 40:  41: mapper.BeforeMapSet += (IModelInspector modelInspector, PropertyPath member, ISetPropertiesMapper propertyCustomizer) => 42: { 43: if (modelInspector.IsManyToMany(member.LocalMember) == true) 44: { 45: propertyCustomizer.Key(x => x.Column(member.LocalMember.DeclaringType.Name + "_Id")); 46:  47: Type sourceType = member.LocalMember.DeclaringType; 48: Type destinationType = member.LocalMember.GetPropertyOrFieldType().GetGenericArguments().First(); 49: IEnumerable<String> names = new Type[] { sourceType, destinationType }.Select(x => x.Name).OrderBy(x => x); 50:  51: //set inverse on the relation of the alphabetically first entity name 52: propertyCustomizer.Inverse(sourceType.Name == names.First()); 53: //set mapping table name from the entity names in alphabetical order 54: propertyCustomizer.Table(String.Join("_", names)); 55: } 56: }; We have to understand how the conventions mapper thinks: For each collection of entities found, it will ask the mapper if it is a one-to-many; in our case, if the collection is a generic one that has an entity as its generic parameter, and the generic parameter type has a similar collection, then it is not a one-to-many; Next, the mapper will ask if the collection that it now knows is not a one-to-many is a many-to-many; Before a set is mapped, if it corresponds to a many-to-many, we set its mapping table. Now, this is tricky: because we have no way to maintain state, we sort the names of the two endpoint entities and we combine them with a “_”; for the first alphabetical entity, we set its relation to inverse – remember, on a many-to-many relation, only one endpoint must be marked as inverse; finally, we set the column name as the name of the entity with an “_Id” suffix; Before the many-to-many relation is processed, we set the column name as the name of the other endpoint entity with the “_Id” suffix, as we did for the set. And that’s it. With these rules, NHibernate will now happily find and configure many-to-many relations, as well as all the others. You can wrap this in a new conventions mapper class, so that it is more easily reusable: 1: public class ManyToManyConventionModelMapper : ConventionModelMapper 2: { 3: public ManyToManyConventionModelMapper() 4: { 5: base.IsOneToMany((MemberInfo member, Boolean isLikely) => 6: { 7: return (this.IsOneToMany(member, isLikely)); 8: }); 9:  10: base.IsManyToMany((MemberInfo member, Boolean isLikely) => 11: { 12: return (this.IsManyToMany(member, isLikely)); 13: }); 14:  15: base.BeforeMapManyToMany += this.BeforeMapManyToMany; 16: base.BeforeMapSet += this.BeforeMapSet; 17: } 18:  19: protected virtual Boolean IsManyToMany(MemberInfo member, Boolean isLikely) 20: { 21: //a relation is many to many if it isn't one to many 22: Boolean isOneToMany = this.ModelInspector.IsOneToMany(member); 23: return (!isOneToMany); 24: } 25:  26: protected virtual Boolean IsOneToMany(MemberInfo member, Boolean isLikely) 27: { 28: Type sourceType = member.DeclaringType; 29: Type destinationType = member.GetMemberFromDeclaringType().GetPropertyOrFieldType(); 30:  31: //check if the property is of a generic collection type 32: if ((destinationType.IsGenericCollection() == true) && (destinationType.GetGenericArguments().Length == 1)) 33: { 34: Type destinationEntityType = destinationType.GetGenericArguments().Single(); 35:  36: //check if the type of the generic collection property is an entity 37: if (this.ModelInspector.IsEntity(destinationEntityType) == true) 38: { 39: //check if there is an equivalent property on the target type that is also a generic collection and points to this entity 40: PropertyInfo collectionInDestinationType = destinationEntityType.GetProperties().Where(x => (x.PropertyType.IsGenericCollection() == true) && (x.PropertyType.GetGenericArguments().Length == 1) && (x.PropertyType.GetGenericArguments().Single() == sourceType)).SingleOrDefault(); 41:  42: if (collectionInDestinationType != null) 43: { 44: return (false); 45: } 46: } 47: } 48:  49: return (true); 50: } 51:  52: protected virtual new void BeforeMapManyToMany(IModelInspector modelInspector, PropertyPath member, IManyToManyMapper collectionRelationManyToManyCustomizer) 53: { 54: Type destinationEntityType = member.LocalMember.GetPropertyOrFieldType().GetGenericArguments().First(); 55: //set the mapping table column names from each source entity name plus the _Id sufix 56: collectionRelationManyToManyCustomizer.Column(destinationEntityType.Name + "_Id"); 57: } 58:  59: protected virtual new void BeforeMapSet(IModelInspector modelInspector, PropertyPath member, ISetPropertiesMapper propertyCustomizer) 60: { 61: if (modelInspector.IsManyToMany(member.LocalMember) == true) 62: { 63: propertyCustomizer.Key(x => x.Column(member.LocalMember.DeclaringType.Name + "_Id")); 64:  65: Type sourceType = member.LocalMember.DeclaringType; 66: Type destinationType = member.LocalMember.GetPropertyOrFieldType().GetGenericArguments().First(); 67: IEnumerable<String> names = new Type[] { sourceType, destinationType }.Select(x => x.Name).OrderBy(x => x); 68:  69: //set inverse on the relation of the alphabetically first entity name 70: propertyCustomizer.Inverse(sourceType.Name == names.First()); 71: //set mapping table name from the entity names in alphabetical order 72: propertyCustomizer.Table(String.Join("_", names)); 73: } 74: } 75: } Conclusion Of course, there is much more to mapping than this, I suggest you look at all the events and functions offered by the ModelMapper to see where you can hook for making it behave the way you want. If you need any help, just let me know!

    Read the article

  • Java Spotlight Episode 149: Geertjan Wielenga on NetBeans 7.4 @netbeans @geertjanw

    - by Roger Brinkley
    Interview with Geertjan Wielenga on the NetBeans 7.4 release Right-click or Control-click to download this MP3 file. You can also subscribe to the Java Spotlight Podcast Feed to get the latest podcast automatically. If you use iTunes you can open iTunes and subscribe with this link: Java Spotlight Podcast in iTunes. Show Notes News Everything in JavaFX is open sourced - Media is now opensource openjdk email Java SE 7 Update 45 Released 7u45 Caller-Allowable-Codebase and Trusted-Library Updated Security Baseline (7u45) impacts Java 7u40 and before with High Security settings What to do if your applet is blocked or warns of “mixed code”? LiveConnect changes in 7u45 JDK 8 end-game details and proposal for making a few exceptions Events Oct 28-30, JAX London, London Nov 4-8, Oredev, Malmö, Sweden Nov 6, JFall, Amsterdam, Netherlands Nov 11-15, Devoxx, Belgium Feature Interview Geertjan Wielenga is a principal product manager in Oracle for NetBeans and has been a member of the NetBeans Team for the past 7 years. NetBeans twitter: https://twitter.com/netbeans Geertjan’s twitter: https://twitter.com/geertjanw The Top Ten Coolest Features in NetBeans IDE 7.4 What’s Cool Nighthacking with James Gosling Arun Gupta waves goodbye and says hello JavaOne 2013 Roundup: Java 8 is Revolutionary, Java is back Survey on Java erasure/reification

    Read the article

  • Oracle Customer Experience Summit @ OpenWorld

    - by Michael Seback
    Businesses worldwide are operating in a new era. Customers are taking charge of their relationships with brands, and the customer experience has become the most important differentiator and driver of business value. Where is the experience heading? And how can businesses take advantage of the customer experience revolution?  Find out from experts at a one-of-a-kind event:  Oracle Customer Experience Summit @ OpenWorld Preview the Conference Schedule for October 3 – 5, 2012 Registration - Wednesday October 3, 7:00 a.m.–6:30 p.m. Westin St. Francis, Moscone West, South, Hilton San Francisco, and Hotel Nikko Sample Sessions: The Experience Imperative - Wednesday October 3, 12:30 p.m.–2:30 p.m. Mark Hurd, President, Oracle Anthony Lye, Senior Vice President, Oracle Cloud Applications Strategy David Vap, Global Vice President, Product Development, Oracle Mike Svatek, Chief Strategy Officer, Bazaarvoice Leading the Experience Revolution - Wednesday October 3, 3:45 p.m.–4:45 p.m. Seth Godin, Best-Selling Author, Founder of Squidoo.com David Vap, Global Vice President, Product Development, Oracle Driving a Customer Experience Strategy - Wednesday October 3, 5:00 p.m.–6:00 p.m. David Vap, Global Vice President, Product Development, Oracle Matthew Banks, Senior Director, Customer Experience Solutions, Oracle Register now.

    Read the article

  • Problem with SNMP and MIBs

    - by jap1968
    I am installing Zabbix to monitor via snmp some devices from a machine running Ubuntu 12.04 server. There is a problem with MIB definitions, since snmp commands do no properly translate some of the MIBs. I have already installed the "snmp-mibs-downloader" package, so the files containing the MIB descriptions are properly installed. The MIB are only translated to obtain the numeric key (the MIB files are accessible to the snmp commands), but the results returned by the snmpget command do not properly translate the key. The zabbix templates that I am using do expect the key translated (SNMPv2-MIB::sysUpTime.0) , so, the current results are not recognised and these are ignored. Test case: $ snmptranslate -On SNMPv2-MIB::sysUpTime.0 .1.3.6.1.2.1.1.3.0 $ snmpget -v 2c -c public 192.168.1.1 1.3.6.1.2.1.1.3.0 iso.3.6.1.2.1.1.3.0 = Timeticks: (2911822510) 337 days, 0:23:45.10 On another machine (running a very old Red Hat based distribution), the snmp commands perform both, the directe and reverse traslation, as expected: # snmptranslate -On SNMPv2-MIB::sysUpTime.0 .1.3.6.1.2.1.1.3.0 # snmpget -v 2c -c public 192.168.1.1 1.3.6.1.2.1.1.3.0 SNMPv2-MIB::sysUpTime.0 = Timeticks: (2911819485) 337 days, 0:23:14.85 What is the problem on my Ubuntu box? Is there something I am missing?

    Read the article

  • Practical Approaches to increasing Virtualization Density-Part 1

    - by Girish Venkat
    Happy New year everyone!. Let me kick start the year off by talking about Virtualization density.  What is it?The number of virtual servers that a physical server can support and it's increase from the prior physical infrastructure as a percentage. Why is it important?This is important because the density should be indicative of how well the server is getting consumed?So what is wrong ?Virtualization density fails to convey the "Real usage" of a server.  Most of the hypervisor based O/S Virtualization  evangelists take pride in the fact that they are now running a Virtual Server farm of X machines compared to a Physical server farm of Y (with Y less than X obviously). The real question is - has your utilization of the server really increased or not.  In an internal study that was conducted by one of the top financial institution - the utilization of servers only went up by 15% from 30 to 45. So, this really means that just by increasing virtualization density one will not be achieving the goal of using up the servers in their server farm better.  I will write about what the possible approaches are to increase virtualization density in the next entry. 

    Read the article

  • SQL SERVER – Difference between COUNT(DISTINCT) vs COUNT(ALL)

    - by pinaldave
    This blog post is written in response to the T-SQL Tuesday hosted by Jes Schultz Borland. Earlier today, I was presenting a 45-minute session at the Community College about “The Beginning SQL Server Database”. One of the students asked me the following question. What is the difference between COUNT(DISTINCT) vs COUNT(ALL)? I found this question from the student very interesting. He seems to have read the documentation (Book Online) and was then asking me this question. I always carry laptop which has SQL Server installed. I quickly opened it and ran the following script. After looking at the result, I think it was clear to everybody. Here is the script: SELECT COUNT([Title]) Value FROM [AdventureWorks].[Person].[Contact] GO SELECT COUNT(ALL [Title]) ALLValue FROM [AdventureWorks].[Person].[Contact] GO SELECT COUNT(DISTINCT [Title]) DistinctValue FROM [AdventureWorks].[Person].[Contact] GO The above script will give me the following results. You can clearly notice from the result set that COUNT (ALL ColumnName) is the same as COUNT(ColumnName). The reality is that the “ALL” is actually  the default option and it needs not to be specified. The ALL keyword includes all the non-NULL values. I know this is very simple and may be it does not change how we work; however looking at the whole angle, I really enjoyed the question. Reference: Pinal Dave (http://blog.SQLAuthority.com) Filed under: Pinal Dave, PostADay, SQL, SQL Authority, SQL Query, SQL Scripts, SQL Server, SQL Tips and Tricks, SQLAuthority News, T SQL, Technology

    Read the article

  • HOUG Konferencia 2012

    - by user645740
    2012. március 26-án, azaz ma kezdodik Egerszalókon a magyarországi Oracle felhasználók rendszeres konferenciája, a HOUG Konferencia 2012, rengeteg érdekes eloadással, kerekasztal beszélgetéssel, networkinggel. A helyszínen még lehet személyesen regisztrálni, a HOUG tagoknak kedvezményes a részvétel. www.houg.hu Igen nagy az érdeklodés! A szálloda és néhány közeli szálláshely már megtelt. A környéken szerencsére még sok szálloda és fogadó található, tehát aki akar még mindig talál szobát. A programról, ami metekintheto, letöltheto: http://www.houg.hu/contents/files/houg2012_program_vegleges.pdf Kedden, március 27-én 10:15-11h John Abel fog izgalmas eloadást tartani az Oracle intelligens tervezett célrendszerekrol - "Extreme Performance with Oracle Engineered Systems", John Abel, Business Development director - EMEA, köztük az Oracle Exadata Database Machine adatázisgéprol is. Szintén kedden délután kerekasztal beszélgetésen fogjuk megtárgyalni ezen rendszerek magyarországi alkalmazásának lehetoségeit, igényeket: Exadata, SPARC SuperCluster, Exalogic, Exalytics, stb. 16:35-17:40, "Célrendszerek alkalmazhatósága a mindennapokban", a kerekasztal beszélgetés moderátora Fodor Endre. Kedd délután eloadást tartok az 13:45-14:35, "Oracle Exadata Database Machine: a világ leggyorsabb adatbázisgépe" címmel. A kedd délutáni Adatgyujtéstol az elemzésig szekció az üzleti intelligencia és adattárház ügyfél tapasztalatokról és technológiákról szól. A szekció vezetoi: Arató Bence, HOUG és Fekete Zoltán (én). BMW tesztvezetés, színházi és koncertprogram színesíti a rendezvényt.

    Read the article

  • MySQL Cluster 7.3 - Join This Week's Webinar to Learn What's New

    - by Mat Keep
    The first Development Milestone and Early Access releases of MySQL Cluster 7.3 were announced just several weeks ago. To provide more detail and demonstrate the new features, Andrew Morgan and I will be hosting a live webinar this coming Thursday 25th October at 0900 Pacific Time / 16.00 UTC Even if you can't make the live webinar, it is still worth registering for the event as you will receive a notification when the replay will be available, to view on-demand at your convenience In the webinar, we will discuss the enhancements being previewed as part of MySQL Cluster 7.3, including: - Foreign Key Constraints: Yes, we've looked into the future and decided Foreign Keys are it ;-) You can read more about the implementation of Foreign Keys in MySQL Cluster 7.3 here - Node.js NoSQL API: Allowing web, mobile and cloud services to query and receive results sets from MySQL Cluster, natively in JavaScript, enables developers to seamlessly couple high performance, distributed applications with a high performance, distributed, persistence layer delivering 99.999% availability. You can study the Node.js / MySQL Cluster tutorial here - Auto-Installer: This new web-based GUI makes it simple for DevOps teams to quickly configure and provision highly optimized MySQL Cluster deployments on-premise or in the cloud You can view a YouTube tutorial on the MySQL Cluster Auto-Installer here  So we have a lot to cover in our 45 minute session. It will be time well spent if you want to know more about the future direction of MySQL Cluster and how it can help you innovate faster, with greater simplicity. Registration is open 

    Read the article

  • How do i delete these files?

    - by user107277
    I ran this command sudo find / -type d -name '*Trash*' | sudo xargs du -h | sort This was the output: 100M /root/.local/share/Trash/files/recup_dir.2.30 100M /root/.local/share/Trash/files/recup_dir.2.72 100M /root/.local/share/Trash/files/recup_dir.32 101M /root/.local/share/Trash/files/recup_dir.2.27 101M /root/.local/share/Trash/files/recup_dir.29 103M /root/.local/share/Trash/files/recup_dir.2.7 103M /root/.local/share/Trash/files/recup_dir.9 103M /root/.local/share/Trash/files/recup_dir.93 106M /root/.local/share/Trash/files/recup_dir.187 106M /root/.local/share/Trash/files/recup_dir.71 107M /root/.local/share/Trash/files/recup_dir.131 107M /root/.local/share/Trash/files/recup_dir.136 107M /root/.local/share/Trash/files/recup_dir.2.46 107M /root/.local/share/Trash/files/recup_dir.51 108M /root/.local/share/Trash/files/recup_dir.106 108M /root/.local/share/Trash/files/recup_dir.2.78 108M /root/.local/share/Trash/files/recup_dir.52 109M /root/.local/share/Trash/files/recup_dir.2.32 109M /root/.local/share/Trash/files/recup_dir.34 110M /root/.local/share/Trash/files/recup_dir.2.28 110M /root/.local/share/Trash/files/recup_dir.2.53 110M /root/.local/share/Trash/files/recup_dir.30 110M /root/.local/share/Trash/files/recup_dir.55 110M /root/.local/share/Trash/files/recup_dir.89 112M /root/.local/share/Trash/files/recup_dir.2.31 112M /root/.local/share/Trash/files/recup_dir.33 114M /root/.local/share/Trash/files/recup_dir.2.29 114M /root/.local/share/Trash/files/recup_dir.2.74 114M /root/.local/share/Trash/files/recup_dir.31 115M /root/.local/share/Trash/files/recup_dir.125 117M /root/.local/share/Trash/files/recup_dir.83 118M /root/.local/share/Trash/files/recup_dir.105 118M /root/.local/share/Trash/files/recup_dir.2.70 119M /root/.local/share/Trash/files/recup_dir.133 1.1G /root/.local/share/Trash/files/recup_dir.148 11M /root/.local/share/Trash/files/recup_dir.179 1.1M /root/.local/share/Trash/info 122M /root/.local/share/Trash/files/recup_dir.80 124M /root/.local/share/Trash/files/recup_dir.137 125G /root/.local/share/Trash 125G /root/.local/share/Trash/files 125M /root/.local/share/Trash/files/recup_dir.2.49 129M /root/.local/share/Trash/files/recup_dir.153 1.2G /root/.local/share/Trash/files/recup_dir.165 1.2G /root/.local/share/Trash/files/recup_dir.166 12K /media/A80E1DE60E1DAE76/.Trash-1000/files 12M /root/.local/share/Trash/files/recup_dir.178 12M /root/.local/share/Trash/files/recup_dir.180 12M /root/.local/share/Trash/files/recup_dir.181 130M /root/.local/share/Trash/files/recup_dir.85 137M /root/.local/share/Trash/files/recup_dir.2.5 137M /root/.local/share/Trash/files/recup_dir.7 137M /root/.local/share/Trash/files/recup_dir.76 13M /root/.local/share/Trash/files/recup_dir.143 13M /root/.local/share/Trash/files/recup_dir.18 13M /root/.local/share/Trash/files/recup_dir.182 13M /root/.local/share/Trash/files/recup_dir.2.16 13M /root/.local/share/Trash/files/recup_dir.2.2 13M /root/.local/share/Trash/files/recup_dir.4 140M /root/.local/share/Trash/files/recup_dir.2.77 145M /root/.local/share/Trash/files/recup_dir.2.63 147M /root/.local/share/Trash/files/recup_dir.2.43 147M /root/.local/share/Trash/files/recup_dir.45 148M /root/.local/share/Trash/files/recup_dir.84 149M /root/.local/share/Trash/files/recup_dir.160 149M /root/.local/share/Trash/files/recup_dir.2.79 1.4G /root/.local/share/Trash/files/recup_dir.191 150M /root/.local/share/Trash/files/recup_dir.2.26 150M /root/.local/share/Trash/files/recup_dir.28 153M /root/.local/share/Trash/files/recup_dir.64 153M /root/.local/share/Trash/files/recup_dir.78 154M /root/.local/share/Trash/files/recup_dir.107 155M /root/.local/share/Trash/files/recup_dir.2.80 155M /root/.local/share/Trash/files/recup_dir.79 15M /root/.local/share/Trash/files/recup_dir.151 162M /root/.local/share/Trash/files/recup_dir.65 163M /root/.local/share/Trash/files/recup_dir.82 164M /root/.local/share/Trash/files/recup_dir.104 165M /root/.local/share/Trash/files/recup_dir.2.39 165M /root/.local/share/Trash/files/recup_dir.41 168M /root/.local/share/Trash/files/recup_dir.2.62 16M /root/.local/share/Trash/files/recup_dir.171 170M /root/.local/share/Trash/files/recup_dir.135 170M /root/.local/share/Trash/files/recup_dir.159 171M /root/.local/share/Trash/files/recup_dir.91 172M /root/.local/share/Trash/files/recup_dir.2.41 172M /root/.local/share/Trash/files/recup_dir.43 175M /root/.local/share/Trash/files/recup_dir.2.33 175M /root/.local/share/Trash/files/recup_dir.35 176M /root/.local/share/Trash/files/recup_dir.2.76 179M /root/.local/share/Trash/files/recup_dir.2.38 179M /root/.local/share/Trash/files/recup_dir.40 179M /root/.local/share/Trash/files/recup_dir.61 1.7G /root/.local/share/Trash/files/recup_dir.167 17M /root/.local/share/Trash/files/recup_dir.172 180M /root/.local/share/Trash/files/recup_dir.186 181M /root/.local/share/Trash/files/recup_dir.2.71 182M /root/.local/share/Trash/files/recup_dir.158 183M /root/.local/share/Trash/files/recup_dir.2.59 185M /root/.local/share/Trash/files/recup_dir.123 189M /root/.local/share/Trash/files/recup_dir.92 18M /root/.local/share/Trash/files/recup_dir.142 18M /root/.local/share/Trash/files/recup_dir.149 18M /root/.local/share/Trash/files/recup_dir.150 18M /root/.local/share/Trash/files/recup_dir.152 18M /root/.local/share/Trash/files/recup_dir.173 18M /root/.local/share/Trash/files/recup_dir.177 191M /root/.local/share/Trash/files/recup_dir.147 193M /root/.local/share/Trash/files/recup_dir.102 195M /root/.local/share/Trash/files/recup_dir.73 196M /root/.local/share/Trash/files/recup_dir.94 198M /root/.local/share/Trash/files/recup_dir.2.58 19M /root/.local/share/Trash/files/recup_dir.175 19M /root/.local/share/Trash/files/recup_dir.176 205M /root/.local/share/Trash/files/recup_dir.108 206M /root/.local/share/Trash/files/recup_dir.56 206M /root/.local/share/Trash/files/recup_dir.60 207M /root/.local/share/Trash/files/recup_dir.2.55 209M /root/.local/share/Trash/files/recup_dir.90 2.0G /root/.local/share/Trash/files/recup_dir.190 20K /media/A80E1DE60E1DAE76/.Trash-1000/info 20M /root/.local/share/Trash/files/recup_dir.17 20M /root/.local/share/Trash/files/recup_dir.2.15 210M /root/.local/share/Trash/files/recup_dir.121 211M /root/.local/share/Trash/files/recup_dir.134 212M /root/.local/share/Trash/files/recup_dir.57 21M /root/.local/share/Trash/files/recup_dir.174 223M /root/.local/share/Trash/files/recup_dir.88 225M /root/.local/share/Trash/files/recup_dir.118 230M /root/.local/share/Trash/files/recup_dir.87 232M /root/.local/share/Trash/files/recup_dir.66 235M /root/.local/share/Trash/files/recup_dir.139 236M /root/.local/share/Trash/files/recup_dir.97 238M /root/.local/share/Trash/files/recup_dir.2.54 240M /root/.local/share/Trash/files/recup_dir.163 241M /root/.local/share/Trash/files/recup_dir.126 242M /root/.local/share/Trash/files/recup_dir.2.81 243M /root/.local/share/Trash/files/recup_dir.156 244M /root/.local/share/Trash/files/recup_dir.2.37 244M /root/.local/share/Trash/files/recup_dir.39 248M /root/.local/share/Trash/files/recup_dir.110 249M /root/.local/share/Trash/files/recup_dir.75 256M /root/.local/share/Trash/files/recup_dir.2.73 257M /root/.local/share/Trash/files/recup_dir.2.64 25M /root/.local/share/Trash/files/recup_dir.10 25M /root/.local/share/Trash/files/recup_dir.2.8 262M /root/.local/share/Trash/files/recup_dir.86 266M /root/.local/share/Trash/files/recup_dir.144 27M /root/.local/share/Trash/files/recup_dir.99 282M /root/.local/share/Trash/files/recup_dir.127 29M /root/.local/share/Trash/files/recup_dir.183 29M /root/.local/share/Trash/files/recup_dir.22 29M /root/.local/share/Trash/files/recup_dir.2.20 316M /root/.local/share/Trash/files/recup_dir.124 31M /root/.local/share/Trash/files/recup_dir.2.21 31M /root/.local/share/Trash/files/recup_dir.23 320M /root/.local/share/Trash/files/recup_dir.168 32M /root/.local/share/Trash/files/recup_dir.12 32M /root/.local/share/Trash/files/recup_dir.2.10 334M /root/.local/share/Trash/files/recup_dir.140 338M /root/.local/share/Trash/files/recup_dir.69 33M /root/.local/share/Trash/files/recup_dir.21 33M /root/.local/share/Trash/files/recup_dir.2.19 340M /root/.local/share/Trash/files/recup_dir.2.57 341M /root/.local/share/Trash/files/recup_dir.185 342M /root/.local/share/Trash/files/recup_dir.169 343M /root/.local/share/Trash/files/recup_dir.129 346M /root/.local/share/Trash/files/recup_dir.111 348M /root/.local/share/Trash/files/recup_dir.103 351M /root/.local/share/Trash/files/recup_dir.2.34 351M /root/.local/share/Trash/files/recup_dir.36 352M /root/.local/share/Trash/files/recup_dir.155 358M /root/.local/share/Trash/files/recup_dir.59 36G /root/.local/share/Trash/files/recup_dir.1 36G /root/.local/share/Trash/files/recup_dir.2 36M /root/.local/share/Trash/files/recup_dir.120 36M /root/.local/share/Trash/files/recup_dir.2.24 36M /root/.local/share/Trash/files/recup_dir.2.51 36M /root/.local/share/Trash/files/recup_dir.26 37M /root/.local/share/Trash/files/recup_dir.112 390M /root/.local/share/Trash/files/recup_dir.162 398M /root/.local/share/Trash/files/recup_dir.2.67 39M /root/.local/share/Trash/files/recup_dir.145 401M /root/.local/share/Trash/files/recup_dir.2.52 402M /root/.local/share/Trash/files/recup_dir.54 408M /root/.local/share/Trash/files/recup_dir.2.40 408M /root/.local/share/Trash/files/recup_dir.42 4.0K /home/daniel/.local/share/Trash 40K /media/A80E1DE60E1DAE76/.Trash-1000 41M /root/.local/share/Trash/files/recup_dir.13 41M /root/.local/share/Trash/files/recup_dir.2.11 428M /root/.local/share/Trash/files/recup_dir.2.61 434M /root/.local/share/Trash/files/recup_dir.2.36 434M /root/.local/share/Trash/files/recup_dir.38 43M /root/.local/share/Trash/files/recup_dir.19 43M /root/.local/share/Trash/files/recup_dir.2.17 43M /root/.local/share/Trash/files/recup_dir.53 440M /root/.local/share/Trash/files/recup_dir.157 448M /root/.local/share/Trash/files/recup_dir.2.35 448M /root/.local/share/Trash/files/recup_dir.37 44M /root/.local/share/Trash/files/recup_dir.20 44M /root/.local/share/Trash/files/recup_dir.2.18 454M /root/.local/share/Trash/files/recup_dir.116 47M /root/.local/share/Trash/files/recup_dir.11 47M /root/.local/share/Trash/files/recup_dir.2.9 48M /root/.local/share/Trash/files/recup_dir.2.48 495M /root/.local/share/Trash/files/recup_dir.192 49M /root/.local/share/Trash/files/recup_dir.114 49M /root/.local/share/Trash/files/recup_dir.50 52M /root/.local/share/Trash/files/recup_dir.3 538M /root/.local/share/Trash/files/recup_dir.2.68 53M /root/.local/share/Trash/files/recup_dir.95 54M /root/.local/share/Trash/files/recup_dir.98 551M /root/.local/share/Trash/files/recup_dir.63 57M /root/.local/share/Trash/files/recup_dir.101 5.7M /root/.local/share/Trash/files/recup_dir.119 57M /root/.local/share/Trash/files/recup_dir.14 57M /root/.local/share/Trash/files/recup_dir.2.12 581M /root/.local/share/Trash/files/recup_dir.70 586M /root/.local/share/Trash/files/recup_dir.170 588M /root/.local/share/Trash/files/recup_dir.62 58M /root/.local/share/Trash/files/recup_dir.2.4 58M /root/.local/share/Trash/files/recup_dir.2.42 58M /root/.local/share/Trash/files/recup_dir.44 58M /root/.local/share/Trash/files/recup_dir.6 59M /root/.local/share/Trash/files/recup_dir.2.22 59M /root/.local/share/Trash/files/recup_dir.24 603M /root/.local/share/Trash/files/recup_dir.109 60M /root/.local/share/Trash/files/recup_dir.15 60M /root/.local/share/Trash/files/recup_dir.2.13 619M /root/.local/share/Trash/files/recup_dir.154 61M /root/.local/share/Trash/files/recup_dir.2.23 61M /root/.local/share/Trash/files/recup_dir.25 626M /root/.local/share/Trash/files/recup_dir.138 62M /root/.local/share/Trash/files/recup_dir.2.3 62M /root/.local/share/Trash/files/recup_dir.5 63M /root/.local/share/Trash/files/recup_dir.188 64M /root/.local/share/Trash/files/recup_dir.2.1 65M /root/.local/share/Trash/files/recup_dir.113 65M /root/.local/share/Trash/files/recup_dir.146 69M /root/.local/share/Trash/files/recup_dir.122 701M /root/.local/share/Trash/files/recup_dir.2.60 71M /root/.local/share/Trash/files/recup_dir.130 71M /root/.local/share/Trash/files/recup_dir.141 72M /root/.local/share/Trash/files/recup_dir.132 72M /root/.local/share/Trash/files/recup_dir.2.47 74M /root/.local/share/Trash/files/recup_dir.16 74M /root/.local/share/Trash/files/recup_dir.2.14 74M /root/.local/share/Trash/files/recup_dir.2.25 74M /root/.local/share/Trash/files/recup_dir.2.45 74M /root/.local/share/Trash/files/recup_dir.27 74M /root/.local/share/Trash/files/recup_dir.47 751M /root/.local/share/Trash/files/recup_dir.164 752M /root/.local/share/Trash/files/recup_dir.128 76M /root/.local/share/Trash/files/recup_dir.49 77M /root/.local/share/Trash/files/recup_dir.115 77M /root/.local/share/Trash/files/recup_dir.77 8.0K /media/A80E1DE60E1DAE76/.Trash-1000/expunged 810M /root/.local/share/Trash/files/recup_dir.58 815M /root/.local/share/Trash/files/recup_dir.2.66 818M /root/.local/share/Trash/files/recup_dir.2.56 82M /root/.local/share/Trash/files/recup_dir.2.44 82M /root/.local/share/Trash/files/recup_dir.46 835M /root/.local/share/Trash/files/recup_dir.68 84M /root/.local/share/Trash/files/recup_dir.189 860M /root/.local/share/Trash/files/recup_dir.161 86M /root/.local/share/Trash/files/recup_dir.117 86M /root/.local/share/Trash/files/recup_dir.2.69 86M /root/.local/share/Trash/files/recup_dir.2.75 90M /root/.local/share/Trash/files/recup_dir.74 924M /root/.local/share/Trash/files/recup_dir.184 94M /root/.local/share/Trash/files/recup_dir.81 95M /root/.local/share/Trash/files/recup_dir.100 96M /root/.local/share/Trash/files/recup_dir.2.6 96M /root/.local/share/Trash/files/recup_dir.2.65 96M /root/.local/share/Trash/files/recup_dir.8 97M /root/.local/share/Trash/files/recup_dir.2.50 97M /root/.local/share/Trash/files/recup_dir.67 97M /root/.local/share/Trash/files/recup_dir.72 98M /root/.local/share/Trash/files/recup_dir.96 99M /root/.local/share/Trash/files/recup_dir.48 How do I delete these files?

    Read the article

< Previous Page | 95 96 97 98 99 100 101 102 103 104 105 106  | Next Page >