1. 17 Aug, 2017 1 commit
  2. 16 Aug, 2015 1 commit
  3. 09 Aug, 2015 2 commits
  4. 16 Nov, 2014 1 commit
  5. 12 Nov, 2014 1 commit
  6. 06 Nov, 2014 1 commit
  7. 24 Jun, 2014 1 commit
  8. 23 Jun, 2014 1 commit
  9. 06 Feb, 2014 1 commit
  10. 23 Jun, 2011 1 commit
    • Augustin Degomme's avatar
      A lot of testing and feedback on this one is needed, in order to improve the way it is done. · fbf0b6c1
      Augustin Degomme authored
      Summary :
      
      - trace can be dumped on disk while parsing
      - data can be restored, loading only in memory the part of trace we want to display (by time and by containers).
      - a light preview version of the whole trace can be displayed, allowing to chose the interval and actually load data from it
      
      
      How it works :
      
      - allow Serialization of IntervalOfContainers to the disk while parsing. Each finished IntervalOFContainer (containing 10000 states or links) can be dumped to the disk, using Boost serialization library, in a separate file, then its memory is freed, allowing to parse huge files (tested with 8Gb). Each type, container, linked in the IntervalOfContainer is assigned a unique id to avoid serializing too much data (file Serializer.hpp). If Boost with gzip is used, the resulting data is compressed. This is handled by Several SerializerWriter threads, and the assignment to each thread is done by a Singleton object, the SerializerDispatcher. The number of threads used is the number of CPUs found in the machine.
      
      At the end of parsing all remaining loaded intervalOfContainers are dumped. File naming is "Unique ID of the container"_"IntervalOfContainer index in the container". They are saved in a folder named after the trace file, without extension.  
      
      At the end of dumping, we have a folder containing many files. A file called "name of the trace".vite is created in this folder, which handles all containers, types, with their unique IDs. For each IntervalOFContainer of each Container, the beginning and end timings are also saved. This file will be used to correlate data from the multiple IntervalOfContainers files. It stores also the sum of all the times of all StateTypes encountered in each intervalOFContainers.
      
      - we can now open this .vite file. A ParserSplitted is then used, allowing to restore the structure of the trace and all types. 
      
         - If the -t option is specified with a time interval, data is directly reloaded from the serialized files, loading in memory only the intervalofcontainers in the time interval.
         - If the -t option was not specified, we load the preview version of the trace, contained in the .vite file.
      
      The preview version only stores states for the moment. When browsing the preview version, user can select a zone and press ctrl. This opens a new vite window, with the same zoom, but the data is then loaded from the serialized files.
      
      
      
      How to use : 
      
      needed libraries : libboost_serialization, libboost_thread, and libboost_iostreams . These libraries are in the standard boost package. In linux, they include the gzip library needed for compression and bindings. In windows, this library is not included and has to be included after, and boost recompiled.
      
      - cmake : activate the option VITE_ENABLE_SERIALIZATION in order to check for boost libraries, and to add corresponding files 
      - configure :  add the flag --use_boost if libraries are in /usr/lib, --boost_libdir=$dir else.
      - by hand in the src.pro file :  add needed libraries ( -lboost_serialization -lboost_thread -lboost_iostreams ) and flags USE_ITC, BOOST_SERIALIZE, and BOOST_GZIP to activate everything
      
      
      
      todo: 
      - make preview + -c option work together ( -c and -t work together for the moment, -c and preview also, but not when loading actual data from disk)
      - add other data to the preview (links events and variables)
      - check if using lots of threads to compress is really useful
      - better balance between those threads, without rebinding qt signal/slots each time
      - tests, tests and tests.
      - documentation and comments.
      fbf0b6c1
  11. 14 Jun, 2011 1 commit
  12. 11 Jun, 2011 2 commits
    • Augustin Degomme's avatar
      - handle cancel while parsing with multi threaded parser · 790060ff
      Augustin Degomme authored
      - change the way types are searched at each insertion (we iterated through lists at each insertion to find the right types, now we use maps which provide much better searching performance)
      790060ff
    • Augustin Degomme's avatar
      continuing the flood of your mailboxes · e46d22b3
      Augustin Degomme authored
      - a few warnings removed for windows again
      - node selection while zooming now almost keeps the zoom (I don't get why the min changes a little bit, and don't know if it's possible to fix this)
      
      new features:
      -multithreaded Paje parser : this parser uses 3 threads :
         - the parsing thread to read the file and produce lines and tokens (lexical analysis), and aggregate them in blocks of 10000 lines 
         - the builder thread which handles these blocks of lines and calls store_event of the ParserEventPaje and transforms the tokens into the appropriate types, checks the correctness of the line (syntaxic analysis) but doesn't perform the calls to trace and the structural verifications 
        - the trace building thread which performs semantic analysis (if types, containers, exist) and performs the calls to the trace (adding events, states to the trace).
      
      - file mapping : just for multithread version, the file is mapped into memory by chunks (100MB for the moment), which is faster and allows to handle larger files without using too much memory. The limit of 1Gb of the other version is removed, ViTE can now handle much larger files.
      
      
      These features need more testing and feedback and can be activated at compile time by setting the flag MT_PARSING. 
      
      note: the parser still uses a tokens number limit, this will be merged with the new version soon.
      
      
      issues:
      - gracefully stop parsing when cancel button is hit is not handled yet and causes segfaults
      e46d22b3
  13. 12 Nov, 2010 1 commit
  14. 27 Oct, 2010 1 commit
  15. 21 Apr, 2010 1 commit
  16. 03 Mar, 2010 1 commit
  17. 22 Dec, 2009 1 commit
  18. 21 Nov, 2009 1 commit
  19. 24 Aug, 2009 1 commit
  20. 21 Aug, 2009 1 commit
  21. 20 Aug, 2009 1 commit
  22. 18 Aug, 2009 1 commit
  23. 04 Aug, 2009 1 commit
  24. 25 Jul, 2009 2 commits
  25. 23 Jul, 2009 1 commit
  26. 04 Jul, 2009 1 commit
  27. 26 Jun, 2009 1 commit
  28. 25 Jun, 2009 1 commit
  29. 19 Jun, 2009 1 commit
  30. 15 Jun, 2009 1 commit
  31. 13 May, 2009 1 commit
  32. 12 May, 2009 1 commit
  33. 30 Apr, 2009 1 commit
  34. 10 Apr, 2009 1 commit
  35. 09 Apr, 2009 1 commit
  36. 07 Apr, 2009 1 commit
  37. 02 Apr, 2009 1 commit