Mlx

Latest version: v0.13.1

Safety actively analyzes 630305 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 1 of 4

0.13.1

🚀

0.13.0

Highlights
* Block sparse matrix multiply speeds up MoEs by >2x
* [some numbers](https://github.com/ml-explore/mlx/pull/1058)
* Improved quantization algorithm should work well for all networks
* [see evaluations](https://github.com/ml-explore/mlx/pull/1061)
* Improved gpu command submission speeds up training and inference
* [some numbers](https://github.com/ml-explore/mlx/pull/1085)

Core
* Bitwise ops added:
- `mx.bitwise_[or|and|xor]`, `mx.[left|right]_shift`, operator overloads
* Groups added to Conv1d
* Added `mx.metal.device_info` to get better informed memory limits
* Added resettable memory stats
* `mlx.optimizers.clip_grad_norm` and `mlx.utils.tree_reduce` added
* Add `mx.arctan2`
* Unary ops now accept array-like inputs ie one can do `mx.sqrt(2)`

Bugfixes
* Fixed shape for slice update
* Bugfix in quantize that used slightly wrong scales/biases
* Fixed memory leak for multi-output primitives encountered with gradient checkpointing
* Fixed conversion from other frameworks for all datatypes
* Fixed index overflow for matmul with large batch size
* Fixed initialization ordering that occasionally caused segfaults

0.12.2

0.12.0

Highlights

* Faster quantized matmul
* Up to 40% faster QLoRA or prompt processing, [some numbers](https://github.com/ml-explore/mlx/pull/1030#issuecomment-2075606627)

Core

* `mx.synchronize` to wait for computation dispatched with `mx.async_eval`
* `mx.radians` and `mx.degrees`
* `mx.metal.clear_cache` to return to the OS the memory held by MLX as a cache for future allocations
* Change quantization to always represent 0 exactly ([relevant issue](https://github.com/ml-explore/mlx-examples/issues/692))

Bugfixes

* Fixed quantization of a block with all 0s that produced NaNs
* Fixed the `len` field in the buffer protocol implementation

0.11.0

Core

- `mx.block_masked_mm` for block-level sparse matrix multiplication
- Shared events for synchronization and asynchronous evaluation

NN

- `nn.QuantizedEmbedding` layer
- `nn.quantize` for quantizing modules
- `gelu_approx` uses tanh for consistency with PyTorch

0.10.0

Page 1 of 4

© 2024 Safety CLI Cybersecurity Inc. All Rights Reserved.