Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Performance of evaluator #26

Open
tasbolat1 opened this issue Jul 9, 2021 · 0 comments
Open

Performance of evaluator #26

tasbolat1 opened this issue Jul 9, 2021 · 0 comments

Comments

@tasbolat1
Copy link

tasbolat1 commented Jul 9, 2021

Hi @arsalan-mousavian,
I have made some tests on both uploaded models (old version, i call it v1 and ACRONYM) for evaluator model. In my test, i wanted to know how good the evaluator perform for on detecting positive grasps. Using json files within splits folder, I decided evaluate train/test accuracy for positive grasps on box and cylinder categories. For each model/split I run tests three times because data loader has stochastic behavior, thus i provide standard deviation too. The results is shown below:

split Model box accuracy cylinder accuracy
train v1 0.82 (0.22) 0.82 (0.21)
test v1 0.74 (0.24) 0.72 (0.27)
train ACRONYM 0.66 (0.28) 0.66 (0.27)
test ACRONYM 0.54 (0.25) 0.54 (0.25)

Generally I found these results to be poor taking into account 0.5 random choice:

  1. There are substantial overfit between train and test splits
  2. ACRONYM is performing worse than v1 model

Can you please confirm that this results are in line with models? May be i am doing something wrong? Thanks in advance

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant