DiscoverMisreading Chat#138: Distilling the Knowledge in a Neural Network
#138: Distilling the Knowledge in a Neural Network

#138: Distilling the Knowledge in a Neural Network

Update: 2024-09-11
Share

Description

大きなモデルから小さなモデルを作るテクニックを向井が回願しました。ご意見感想などは Reddit やおたより投書箱にお寄せください。iTunes のレビューや星もよろしくね。










<figure class="wp-block-audio"></figure>

















<iframe src="https://docs.google.com/forms/d/e/1FAIpQLSdBvbhI98yeJQV_QWBsl1Q5vY7iohwFN-lJOY2fIh_pfjwRSQ/viewform?embedded=true" frameborder="0" width="100%" height="800" marginheight="0" marginwidth="0" allowfullscreen="true" mozallowfullscreen="true" webkitallowfullscreen="true"></iframe>
Comments 
In Channel
loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

#138: Distilling the Knowledge in a Neural Network

#138: Distilling the Knowledge in a Neural Network

Jun Mukai