Skip to content

arnavg115/agrad

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Agrad - Auto Grad

A substandard autogradient implementation using only Numpy. This is an extension of my ml library and I wish to use it to implement more complex networks. It is an amalgamation of Joel Grus' and Andrej Karpathy's implementation of autograd.

Done:

  • loss: mse, softmax cross-entropy
  • optimizer: basic (SGD), adam, RMSprop, momentum
  • activation/functions: tanh, leaky relu, sigmoid, relu, exp, basic tensor ops (add, subtract, matmul etc.)
  • architectures: Linear (MLP) see mnist.py, Transformers (LLaMA.py)

Todo:

  • building blocks: conv
  • architectures: imagenet, Mamba
  • other: KV cache for transformers, test backprop stability on transformer

About

A substandard autograd implementation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages