Packet 11: Bonus 5

Candidate choices for this parameter are rejected until the Armijo condition is met in a standard implementation of backtracking line search. For 10 points each:
[10m] Name this parameter that varies in adaptive approaches like Adam and RMSprop. In a popular method, xt+1 (“x-sub-t plus one”) equals xt (“x-sub-t”) minus the product of this scalar parameter and a vector-valued function evaluated at xt (“x-sub-t”).
ANSWER: step size [or learning rate; prompt on eta or alpha]
[10e] That descent method is named for this operation, which gives the vector of a function’s partial derivatives.
ANSWER: gradient [or grad; accept gradient descent]
[10h] Adam combines RMSProp with this broad idea, which uses past gradients to inform the size of the current update. Nesterov’s accelerated gradient descent uses a form of this idea based on a look-ahead value of the gradient.
ANSWER: momentum
<TM, Other Science (Math)> | NAFTA-Packet-11

HeardPPBE %M %H %
816.2588%50%25%

Back to bonuses

Conversion

TeamOpponentPart 1Part 2Part 3TotalParts
You Don't Have the CardsStrange Case of Dr. Joker and Mr. Pennywise10101030MEH

Summary

TournamentEditionMatchHeardPPBE %M %H %
2025 NAFTA Online02/14/202647.5075%0%0%
2026 NAFTA at Vanderbilt02/14/2026110.00100%0%0%
2025 NAFTA at Toronto09/13/2025130.00100%100%100%
2025 NAFTA at Maryland09/27/2025120.00100%0%100%
2025 NAFTA at Chicago11/08/2025613.3383%50%0%