Skip to content

Title: Results remain unchanged after code modifications despite prints confirming execution #3

@LIU-coder79

Description

@LIU-coder79

Description
I encountered a behavior where evaluation metrics (IDF1) stopped updating after certain code changes, even when reverting all modifications. The issue persists until the project is fully redeployed.

Steps to Reproduce

  1. Deployed the original code → obtained baseline IDF1: 76.692.
  2. Modified only sam2tracker.py and run.py in several experimental steps — results updated normally initially.
  3. After one specific change, the IDF1 became stuck at 75.318 in all subsequent runs.
  4. Even after removing all modifications (or commenting them out), the result remained 75.318.
  5. Cleared output caches — no effect.
  6. Added print() statements to confirm code execution — prints showed modifications were executed, but the final metric remained unchanged.
  7. Only after completely redeploying the project did the results return to normal.

Expected Behavior
Evaluation results should reflect code changes, especially when modifications are reverted or adjusted.

Current Behavior
IDF1 is fixed at 75.318 regardless of code changes, until a full redeploy is performed.

Information Requested

  • Are there other caches or state files (e.g., in /tmp, ~/.cache, or inside virtual environment) that should be cleared between runs?
  • Any known behavior where certain modules or variables retain state across script executions?

Additional Context
The issue appeared only after several valid iterative changes, suggesting something in the runtime environment or internal state became “stuck.” Full redeployment resolves it, but this is not practical during development.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions