Skip to content
This repository was archived by the owner on Aug 6, 2025. It is now read-only.
This repository was archived by the owner on Aug 6, 2025. It is now read-only.

Hyperparameter tuning - discrepancy between readme and code? #73

@maruker

Description

@maruker

The readme states

For probing tasks, we used an MLP with a Sigmoid nonlinearity and and tuned the nhid (in [50, 100, 200]) and dropout (in [0.0, 0.1, 0.2]) on the dev set.

However, in the code it looks like the parameters given by the user are always used. No tuning takes place and no predefined hyperparameters are loaded. Maybe I missed something?

Should I do hyperparameter tuning to get results that are comparable to the literature?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions