UNPKG

@tensorflow/tfjs-core

Version:

Hardware-accelerated JavaScript library for machine intelligence

131 lines (118 loc) 3.97 kB
/** * @license * Copyright 2018 Google Inc. All Rights Reserved. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * ============================================================================= */ import {customGrad} from '../gradients'; import {Tensor} from '../tensor'; import {GradSaveFunc} from '../tensor_types'; import {convertToTensor} from '../tensor_util_env'; import {TensorLike} from '../types'; import {op} from './operation'; /** * Computes the softmax normalized vector given the logits. * * ```js * const a = tf.tensor1d([1, 2, 3]); * * a.softmax().print(); // or tf.softmax(a) * ``` * * ```js * const a = tf.tensor2d([2, 4, 6, 1, 2, 3], [2, 3]); * * a.softmax().print(); // or tf.softmax(a) * ``` * * @param logits The logits array. * @param dim The dimension softmax would be performed on. Defaults to `-1` * which indicates the last dimension. */ /** @doc {heading: 'Operations', subheading: 'Normalization'} */ function softmax_<T extends Tensor>(logits: T|TensorLike, dim = -1): T { const $logits = convertToTensor(logits, 'logits', 'softmax'); if (dim === -1) { dim = $logits.rank - 1; } if (dim !== $logits.rank - 1) { throw Error( 'Softmax along a non-last dimension is not yet supported. ' + `Logits was rank ${$logits.rank} and dim was ${dim}`); } const customOp = customGrad((logits: Tensor, save: GradSaveFunc) => { // Do it in log space for numerical stability. // exp(X - logSumExp(X)) const keepDims = true; const lse = logits.logSumExp([dim], keepDims); const logResult = logits.toFloat().sub(lse); const y = logResult.exp() as T; save([y]); const gradFunc = (dy: T, saved: Tensor[]) => { const [y] = saved; const dyTimesY = dy.mul(y); const keepDims = true; return dyTimesY.sub(dyTimesY.sum([dim], keepDims).mul(y)); }; return {value: y, gradFunc}; }); return customOp($logits); } /** * Computes the log softmax. * * ```js * const a = tf.tensor1d([1, 2, 3]); * * a.logSoftmax().print(); // or tf.logSoftmax(a) * ``` * * ```js * const a = tf.tensor2d([2, 4, 6, 1, 2, 3], [2, 3]); * * a.logSoftmax().print(); // or tf.logSoftmax(a) * ``` * * @param logits The logits array. * @param axis The dimension softmax would be performed on. Defaults to `-1` * which indicates the last dimension. */ /** @doc {heading: 'Operations', subheading: 'Normalization'} */ function logSoftmax_<T extends Tensor>(logits: T|TensorLike, axis = -1): T { const $logits = convertToTensor(logits, 'logits', 'logSoftmax'); if (axis === -1) { axis = $logits.rank - 1; } if (axis !== $logits.rank - 1) { throw Error( 'Log Softmax along a non-last dimension is not yet supported. ' + `Logits was rank ${$logits.rank} and axis was ${axis}`); } const customOp = customGrad((logits: Tensor, save: GradSaveFunc) => { const keepDims = true; const xMax = logits.max(axis, true); const shifted = logits.sub(xMax); const value = shifted.toFloat().sub(shifted.exp().sum(axis, keepDims).log()) as T; save([value]); const gradFunc = (dy: T, saved: Tensor[]) => { const [value] = saved; const softmax = value.exp(); return dy.sub(dy.sum(axis, keepDims).mul(softmax)); }; return {value, gradFunc}; }); return customOp($logits); } export const softmax = op({softmax_}); export const logSoftmax = op({logSoftmax_});