Index

Modules: data, graph, image, literal, nimxla, nimxla_plot, nn, plots, shape, tensor, train.

API symbols

`!=`:
`!`:
`$`:
`*`:
`+`:
`-`:
`/`:
`<=`:
`<`:
`==`:
`>=`:
`>`:
`@@`:
`[]=`:
`[]`:
`^`:
abs:
accuracyFunc:
AdamOptimizer:
add:
addAt:
addrOf:
Affine:
append:
approxEqual:
argMax:
argMin:
ArrayKind:
arrayShape:
at:
avgPool1d:
avgPool2d:
avgPool3d:
batchNormGrad:
batchNormInference:
batchNormTraining:
BF16:
Bool:
boolean:
broadcast:
broadcastInDim:
Buffer:
build:
Builder:
BuilderError:
C128:
C64:
ceil:
ChainedScheduler:
CIFAR10Dataset:
cifar10Dataset:
clamp:
classes:
Client:
clone:
collapse:
compile:
compileTest:
compileTrain:
Computation:
concat:
constant:
constantInit:
conv1d:
conv2d:
conv3d:
convert:
convolution:
copy:
cos:
CosineAnnealingLR:
crossEntropyLoss:
DataLoader:
Dataset:
DataType:
decomposeTuple:
deviceCount:
dims:
Direction:
dot:
dropout:
dtype:
dtypeOf:
dump:
Elastic:
ElemType:
Error:
errorNode:
Executable:
exp:
F16:
f32:
F32:
f64:
F64:
Fatal:
fill:
flatten:
Flip:
floor:
format:
fromHlo:
gather:
getAccuracy:
getBatch:
getItem:
getLayout:
getParams:
glorotInit:
gradient:
gradNames:
gridLayout:
heInit:
HloModule:
Horizontal:
I16:
i32:
I32:
i64:
I64:
I8:
ImageOp:
ImageOpKind:
ImageRequest:
Info:
init:
initBatchNorm:
initConv2d:
initConv2dBatchNorm:
InitFunc:
initLinear:
initParams:
initRandom:
initTransformer:
InvalidType:
iota:
isFinite:
learnableVars:
learningRate:
len:
LinearLR:
lit:
Literal:
loadCheckpoint:
log:
log1p:
logicalAnd:
logicalOr:
LogLevel:
makeTuple:
max:
maxPool1d:
maxPool2d:
maxPool3d:
maxValue:
mean:
min:
minValue:
MNISTDataset:
mnistDataset:
Module:
mseLoss:
name:
narrow:
newBuffer:
newBuilder:
newChainedScheduler:
newClient:
newCosineAnnealingLR:
newCPUClient:
newGPUClient:
newLinearLR:
newLiteral:
newLoader:
newStepLR:
newTensor:
newTPUClient:
newVariable:
Node:
normalInit:
normalization:
noutputs:
one:
oneHot:
OpaqueType:
openWebSocket:
Opt2d:
Opt3d:
optimAdam:
optimAdamW:
Optimizer:
optimSGD:
OpType:
Outputs:
pad:
Pad2d:
Pad3d:
Padding:
padSame:
param:
parameter:
parameters:
Params:
platformName:
platformVersion:
plotImage:
plotImageGrid:
pow:
raiseError:
randomAffine:
randomElastic:
randomFlip:
randomWrap:
rank:
rawPtr:
readCheckpointFile:
readTensor:
reduce:
reduceWindow:
relu:
rem:
repr:
reshape:
resultShape:
reverse:
rngNormal:
rngUniform:
round:
rsqrt:
run:
runAndUnpack:
runWith:
saveCheckpoint:
scatter:
Scheduler:
select:
selectAndScatter:
seq2:
seq3:
servePlots:
setLearningRate:
setLogLevel:
setParams:
setPrintOpts:
setVars:
SGDOptimizer:
shape:
Shape:
shape:
ShapeKind:
shutdown:
sigmoid:
sign:
sin:
softmax:
sqrt:
start:
statsPlots:
step:
StepLR:
sum:
tAbs:
tAdd:
tAnd:
tanh:
tArgmax:
tArgmin:
tBatchNormGrad:
tBatchNormInference:
tBatchNormTraining:
tBroadcast:
tBroadcastInDim:
tCeil:
tClamp:
tCollapse:
tConcat:
tConst:
tConv:
tConvert:
tCopy:
tCos:
tDiv:
tDot:
Tensor:
tEq:
tError:
testFunc:
tExp:
tFloor:
tGather:
tGe:
tGt:
tIota:
tIsFinite:
tLe:
tLiteral:
tLog:
tLog1p:
tLt:
tMax:
tMaxPool:
tMin:
tMul:
tNarrow:
tNe:
tNeg:
tNone:
tNot:
toHlo:
Token:
toLiteral:
toLiterals:
tOr:
toSeq:
toShape:
toString:
toTensor:
tPad:
tParam:
tPow:
trainEpoch:
Trainer:
trainFunc:
trainNetwork:
TransContext:
transform:
Transformer:
transpose:
tReduce:
tReduceMax:
tReduceMin:
tReduceSum:
tReduceWindow:
tRelu:
tRem:
tReshape:
tReverse:
tRngNormal:
tRngUniform:
tRound:
tRsqrt:
tScatter:
tSelect:
tSelectAndScatter:
tSigmoid:
tSign:
tSin:
tSoftmax:
tSqrt:
tSub:
tSumPool:
tTanh:
tTranspose:
tTuple:
tTupleElement:
Tuple:
tuple2:
tuple3:
tuple4:
tuple5:
TupleKind:
tZerosLike:
U16:
U32:
U64:
U8:
uniformInit:
update:
updatePlot:
Variable:
varNames:
Vertical:
Warning:
Wrap:
write:
zero:
zeros:
zerosLike: