Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How many subnets does knowledge distillation optimize? #54

Open
swapnilsayansaha opened this issue May 14, 2021 · 3 comments
Open

How many subnets does knowledge distillation optimize? #54

swapnilsayansaha opened this issue May 14, 2021 · 3 comments

Comments

@swapnilsayansaha
Copy link

I have a question that is not cleared in the paper. During knowledge distillation, do you optimize for all 10^19 networks? The elastic - nn portion of the code seems to point to that:

	subnet_settings = []
	for d in depth_list:
		for e in expand_ratio_list:
			for k in ks_list:
				for w in width_mult_list:
					for img_size in image_size_list:
						subnet_settings.append([{
							'image_size': img_size,
							'd': d,
							'e': e,
							'ks': k,
							'w': w,
						}, 'R%s-D%s-E%s-K%s-W%s' % (img_size, d, e, k, w)])
@Darshcg
Copy link

Darshcg commented May 17, 2021

Hi @swapnilsayansaha,

As per my understanding, this code is just generating the different Subnet Settings, I do not see any Implementation regarding KD in the Code base.

@swapnilsayansaha
Copy link
Author

@Darshcg how many subnets does it consider? all 10^19?

@Darshcg
Copy link

Darshcg commented May 17, 2021

@swapnilsayansaha I am not exactly sure, but in this, it is creating 8 subnet settings in this code for MobilenetV3.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants