You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
First, I appreciate your fast responses - it is of course not obvious :)
I have a couple of questions about the training process:
What's the difference between the arguments --cpkt and --pretrained_model in train.py? Isn't a pre-trained model just a model that is trained for some epochs, like a checkpoint?
I was wondering if there are blocks of code that are not trained, and used as pretrained models. For example, the VoxelNet part - do we actually train the subnetwork that processes the voxels into the overhead-view pseudo image?
Is the VoxelNet backbone responsible for both voxelizing the point cloud and creating the overhead-view pseudo?
We've trained the model with 10% of the training data for 80 epochs and a batch_size of 4. To our big surprise, it performed almost as good as the model trained on the full data that you referred to here: pre-trained model #9. Does it make any sense?
Do we use RCNNs at the first stage for centerpoint.yaml config?
We're struggling to understand the heatmap concept. How it is created, and how the GaussianFocalLoss loss function is applied to it. Do you have any hint where we might find some answer for beginners? Google assume we've been born with that knowledge.
Thanks :)
The text was updated successfully, but these errors were encountered:
Hey there!
First, I appreciate your fast responses - it is of course not obvious :)
I have a couple of questions about the training process:
What's the difference between the arguments --cpkt and --pretrained_model in train.py? Isn't a pre-trained model just a model that is trained for some epochs, like a checkpoint?
I was wondering if there are blocks of code that are not trained, and used as pretrained models. For example, the VoxelNet part - do we actually train the subnetwork that processes the voxels into the overhead-view pseudo image?
Is the VoxelNet backbone responsible for both voxelizing the point cloud and creating the overhead-view pseudo?
We've trained the model with 10% of the training data for 80 epochs and a batch_size of 4. To our big surprise, it performed almost as good as the model trained on the full data that you referred to here: pre-trained model #9. Does it make any sense?
Do we use RCNNs at the first stage for centerpoint.yaml config?
We're struggling to understand the heatmap concept. How it is created, and how the GaussianFocalLoss loss function is applied to it. Do you have any hint where we might find some answer for beginners? Google assume we've been born with that knowledge.
Thanks :)
The text was updated successfully, but these errors were encountered: