Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

[D] Questions on general research practices

Hi everyone,

I have a few questions on research practices that are generally followed but rarely mentioned in papers.

  1. Let’s say I have a train-dev-test split. After finding the best hyperparameters on dev set, should I retrain the model on train+dev set before evaluating it on the test set? Some discussions say yes, others say it depends on you and how much data you have and some say no.
  2. Let’s say I’m showcasing results on multiple datasets. Can one change the hyperparameters (learning rate, batch size, etc) from one dataset to another? More importantly, can I change, let’s say, number of units in a layer without adding more layers? Would this count as an architectural change?
  3. If yes, how would answer to above question change if the same is done within the dataset itself containing multiple parts?
  4. Are we allowed to change a publically available dataset? For example, removing outliers for a regression problem?

submitted by /u/manrajsinghgrover
[link] [comments]