Skip to content

param and model file structure

nihui edited this page Feb 23, 2018 · 2 revisions

net.param

example

7767517
3 3
Input         input    0 1 data 0=4 1=4 2=1
InnerProduct  ip       1 1 data fc 0=10 1=1 2=80
Softmax       softmax  1 1 fc prob 0=0

overview

[magic]
  • magic number : 7767517
[layer count] [blob count]
  • layer count : count of the layer line follows, should be exactly the count of all layer names
  • blob count : count of all blobs, usually greater than or equals to the layer count

layer line

[layer type] [layer name] [input count] [output count] [input blobs] [output blobs] [layer specific params]
  • layer type : type name, such as Convolution Softmax etc
  • layer name : name of this layer, must be unique among all layer names
  • input count : count of the blobs this layer needs as input
  • output count : count of the blobs this layer produces as output
  • input blobs : name list of all the input blob names, seperated by space, must be unique among input blob names of all layers
  • output blobs : name list of all the output blob names, seperated by space, must be unique among output blob names of all layers
  • layer specific params : key=value pair list, seperated by space

layer param

0=1 1=2.5 -23303=2,2.0,3.0

key index should be unique in each layer line, pair can be omitted if the default value used

the meaning of existing param key index can be looked up at operation-param-weight-table

  • integer or float key : index 0 ~ 19
  • integer value : int
  • float value : float
  • integer array or float array key : -23300 minus index 0 ~ 19
  • integer array value : [array size],int,int,...,int
  • float array value : [array size],float,float,...,float

net.bin

  +---------+---------+---------+---------+---------+---------+
  | weight1 | weight2 | weight3 | weight4 | ....... | weightN |
  +---------+---------+---------+---------+---------+---------+
  ^         ^         ^         ^
  0x0      0x80      0x140     0x1C0

the model binary is the concatenation of all weight data, each weight buffer is aligned by 32bit

weight buffer

[flag] (optional)
[raw data]
[padding] (optional)
  • flag : unsigned int, little-endian, indicating the weight storage type, 0 => float32, 0x01306B47 => float16, otherwise => quantized int8, may be omitted if the layer implementation forced the storage type explicitly
  • raw data : raw weight data, little-endian, float32 data or float16 data or quantized table and indexes depending on the storage type flag
  • padding : padding space for 32bit alignment, may be omitted if already aligned
You can’t perform that action at this time.