| // Copyright (c) 2018 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| syntax = "proto2"; |
| |
| option optimize_for = LITE_RUNTIME; |
| |
| package assist_ranker; |
| |
| // The weights and biases for a single quantized neural-network layer. All |
| // weight and bias values are 8-bit and can be converted to floats using |
| // value = x * (high - low) / 256 + low. |
| message QuantizedNNLayer { |
| // The weights for the layer. |
| repeated bytes weights = 1; |
| |
| // The bias vectors for the layer. |
| optional bytes biases = 2; |
| |
| // The low value used to dequantize the weights. |
| optional float low = 3; |
| |
| // The high value used to dequantize the weights. |
| optional float high = 4; |
| } |
| |
| // Defines the model weights and biases for a single layer neural network. |
| message QuantizedNNClassifierModel { |
| optional QuantizedNNLayer hidden_layer = 1; |
| optional QuantizedNNLayer logits_layer = 2; |
| } |