File tree 2 files changed +7
-6
lines changed
2 files changed +7
-6
lines changed Original file line number Diff line number Diff line change @@ -222,8 +222,8 @@ class Variable {
222
222
223
223
/* *
224
224
* Run backward pass on the Variable. Gradient of all the inputs
225
- * in the computation graph leading up to the Variable on which the function is
226
- * computed.
225
+ * in the computation graph leading up to the Variable on which the function
226
+ * is computed.
227
227
* @param[in] grad gradient w.r.t to the Variable
228
228
* @param[in] retainGraph If False, clears the input Variables stored
229
229
* by the Variable
@@ -232,8 +232,9 @@ class Variable {
232
232
233
233
/* *
234
234
* Run backward pass on the Variable. Gradient of all the inputs
235
- * in the computation graph leading up to the Variable on which the function is
236
- * computed. Gradient w.r.t the all the elements in the variable is set to 1.0
235
+ * in the computation graph leading up to the Variable on which the function
236
+ * is computed. Gradient w.r.t the all the elements in the variable is set
237
+ * to 1.0
237
238
* @param[in] retainGraph If False, clears the input Variables stored
238
239
* by the Variable
239
240
*/
Original file line number Diff line number Diff line change @@ -65,8 +65,8 @@ class AdaptiveSoftMaxLoss : public BinaryModule {
65
65
* first tail bucket will contain `50 - 5 = 45` targets (subtracting the size
66
66
* of the head bucket), the second tail bucket will contain `100 - 50 = 50`
67
67
* targets (subtracting the size of the first tail bucket). Cutoffs must be
68
- * specified to accommodate all targets: any remaining targets are not assigned
69
- * to an 'overflow' bucket.
68
+ * specified to accommodate all targets: any remaining targets are not
69
+ * assigned to an 'overflow' bucket.
70
70
* @param div_value determines the number of hidden units in the intermediate
71
71
* layer for each tail bucket:
72
72
* \f[
You can’t perform that action at this time.
0 commit comments