Skip to content

GitLab

  • Menu
Projects Groups Snippets
    • Loading...
  • Help
    • Help
    • Support
    • Community forum
    • Submit feedback
    • Contribute to GitLab
  • Sign in
  • Trac Trac
  • Project information
    • Project information
    • Activity
    • Labels
    • Members
  • Issues 246
    • Issues 246
    • List
    • Boards
    • Service Desk
    • Milestones
  • Monitor
    • Monitor
    • Metrics
    • Incidents
  • Analytics
    • Analytics
    • Value stream
  • Wiki
    • Wiki
  • Activity
  • Create a new issue
  • Issue Boards
Collapse sidebar
  • Legacy
  • TracTrac
  • Issues
  • #32890

Closed (moved)
(moved)
Open
Created Jan 07, 2020 by Karsten Loesing@karsten

Remember processed files between module runs

The three recently added modules to archive Snowflake statistics, bridge pool assignments, and BridgeDB metrics have in common that they process any input files regardless of whether they already processed them before.

The problem is that the input files processed by these modules are either never removed (Snowflake statistics) or only removed manually by the operator (bridge pool assignments and BridgeDB statistics).

The effect is that non-recent BridgeDB metrics and bridge pool assignments are being placed in the indexed/recent/ directory in the next execution after they are deleted for being older than 72 hours. The same would happen with Snowflake statistics after the operator removes them from the out/ directory.

The fix is to use a state file containing file names of previously processed files and only process a file not found in there. This is the same approach as taken for bridge descriptor tarballs.

To upload designs, you'll need to enable LFS and have an admin enable hashed storage. More information
Assignee
Assign to
Time tracking