Issue for tracking and coordinating mlx backend work:
mlx.math
- [ ]
fft
- [ ]
fft2
- [ ]
rfft
- [ ]
irfft
- [ ]
stft
- [ ]
istft
- [x]
logsumexp
#19578 - [ ]
qr
- [ ]
segment_sum
#19652 - [ ]
segment_max
#19652 - [x]
erfinv
#19628
mlx.numpy
- [ ]
einsum
- [ ]
bincount
- [ ]
nonzero
- [ ]
cross
- [ ]
vdot
- [ ]
nan_to_num
- [ ]
copy
- [ ]
roll
- [x]
median
#19568 #19574 - [x]
meshgrid
#19574 - [x]
conjugate
- [x]
arctan2
#19759 - [ ]
quantile
- [ ]
imag
- [ ]
real
- [ ]
select
- [x]
argpartition
https://github.com/keras-team/keras/pull/19680 - [ ]
slogdet
- [ ]
select
- [ ]
vectorize
- [ ]
correlate
- [x]
diag
#19714 - [x]
diagonal
#19714
mlx.image
- [x]
rgb_to_grayscale
#19609 - [x]
resize
- #19699
mlx.nn
- [ ]
max_pool
- [ ]
avg_pool
- [ ]
conv
- [ ]
depthwise_conv
- [ ]
separable_conv
- [ ]
conv_transpose
- [ ]
ctc_loss
mlx.rnn
- [ ]
rnn
- [ ]
lstm
- [ ]
gru
mlx.linalg
- [x]
cholesky
- [ ]
det
- [ ]
eig
- [ ]
eigh
- [x]
inv
- [ ]
lu_factor
- [x]
norm
#19698 - [x]
qr
- [ ]
solve
- [ ]
solve_triangular
- [x]
svd
mlx.core
- [x] np.ndarray of i64 is being cast to i32 in mlx during conversion if dtype is not passed
- [x] https://github.com/ml-explore/mlx/issues/1076
- [ ] https://github.com/ml-explore/mlx/issues/1075
- [x] https://github.com/ml-explore/mlx/issues/1066
- [x] https://github.com/ml-explore/mlx/issues/1065
Comment From: lkarthee
PyTest Output
=========================================================================== test session starts ============================================================================
platform darwin -- Python 3.12.2, pytest-8.1.1, pluggy-1.4.0 -- /Users/kartheek/erlang-ws/github-ws/latest/keras/.venv/bin/python3.12
cachedir: .pytest_cache
rootdir: /Users/kartheek/erlang-ws/github-ws/latest/keras
configfile: pyproject.toml
plugins: cov-5.0.0
collected 6 items
keras/src/ops/operation_test.py::OperationTest::test_autoconfig PASSED [ 16%]
keras/src/ops/operation_test.py::OperationTest::test_eager_call PASSED [ 33%]
keras/src/ops/operation_test.py::OperationTest::test_input_conversion FAILED [ 50%]
keras/src/ops/operation_test.py::OperationTest::test_serialization PASSED [ 66%]
keras/src/ops/operation_test.py::OperationTest::test_symbolic_call PASSED [ 83%]
keras/src/ops/operation_test.py::OperationTest::test_valid_naming PASSED [100%]
================================================================================= FAILURES =================================================================================
___________________________________________________________________ OperationTest.test_input_conversion ____________________________________________________________________
self = <keras.src.ops.operation_test.OperationTest testMethod=test_input_conversion>
def test_input_conversion(self):
x = np.ones((2,))
y = np.ones((2,))
z = knp.ones((2,)) # mix
if backend.backend() == "torch":
z = z.cpu()
op = OpWithMultipleInputs()
> out = op(x, y, z)
keras/src/ops/operation_test.py:152:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
keras/src/utils/traceback_utils.py:113: in error_handler
return fn(*args, **kwargs)
keras/src/ops/operation.py:56: in __call__
return self.call(*args, **kwargs)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <Operation name=op_with_multiple_inputs>, x = array([1., 1.]), y = array([1., 1.])
z = <[ValueError('item can only be called on arrays of size 1.') raised in repr()] array object at 0x13f7450c0>
def call(self, x, y, z=None):
# `z` has to be put first due to the order of operations issue with
# torch backend.
> return 3 * z + x + 2 * y
E ValueError: Cannot perform addition on an mlx.core.array and ndarray
keras/src/ops/operation_test.py:14: ValueError
========================================================================= short test summary info ==========================================================================
FAILED keras/src/ops/operation_test.py::OperationTest::test_input_conversion - ValueError: Cannot perform addition on an mlx.core.array and ndarray
======================================================================= 1 failed, 5 passed in 0.13s ========================================================================
How to fix this test case any idea ? add(mx_array, numpy_array) works but fails when using + operator. Should we skip this test for mlx backend ?
Comment From: fchollet
How to fix this test case any idea ? add(mx_array, numpy_array) works but fails when using + operator. Should we skip this test for mlx backend ?
It's not fixable on our side, we should file an issue with the MLX repo. +
will hit array.__add__
which is on their side.
Comment From: Faisal-Alsrheed
Thank you for the list.
I am doing
keras/backend/mlx/nn.py:conv keras/backend/mlx/nn.py:depthwise_conv keras/backend/mlx/nn.py:separable_conv keras/backend/mlx/nn.py:conv_transpose
Comment From: lkarthee
I am working on segment_sum, segment_max, max_pool and avg_pool. Thank you .
Comment From: yrahul3910
I want to take a stab at arctan2
(first-time contributor, so I want to start small). I'm working with the mlx team to see if I can add in the required stuff there first, and then I'll add the implementation here.
Comment From: lkarthee
Thank you @yrahul3910 , please go ahead with adding arctan2
impl.
Comment From: lkarthee
mx.matmul
and mx.tensordot
works only for bfloat16, float16, float32.
FAILED keras/src/ops/numpy_test.py::NumpyDtypeTest::test_tensordot_('int16', 'bool') - ValueError: [matmul] Only real floating point types are supported but int16 and bool were provided which results in int16, which is not a real floating point type.
@fchollet How do we handle this - we can cast integers arguments to float32 if both are integers and result will be float32. If we go this route, we have to modify test cases in numpy_test.py
for mlx. Do you have any suggestions.
Comment From: awni
Just want to let you all know some updates to MLX as of 0.16.1 that may be useful here:
mx.einsum
mx.nan_to_num
mx.conjugate
Are there any high priority items we can fix or add to help move this along?
Comment From: lkarthee
Thank you @awni , we need some help in moving this forward. I will make a list and get back to you in a day or two.
Comment From: acsweet
I'd like to pick up on this issue (first time contributor) starting with fft
if that's okay
Comment From: acsweet
I'm going to start with the "easy" stuff already implemented in mlx, and I'll start in mlx.math
with
- fft2
- rfft
- irfft
- qr
(I'll have to see how to handle the mode
argument from Keras
Comment From: awni
Sounds great! Let us know how we can help on the MLX side.
Comment From: acsweet
@awni Thank you! I'll keep you updated as I progress.
Right now, would it be possible to get stft
and istft
implemented on the mlx side? It looks like it was started here ml-explore/mlx#1004
I saw this implementation too (without an inverse) https://github.com/nuniz/mlx_stft
Comment From: fchollet
Please note, the nn and rnn namespaces are the most important for getting mlx to work with typical workflows.
On Fri, Jan 17, 2025, 4:23 PM acsweet @.***> wrote:
@awni https://github.com/awni Thank you! I'll keep you updated as I progress.
Right now, would it be possible to get stft and istft implemented on the mlx side? It looks like it was started here ml-explore/mlx#1004 https://github.com/ml-explore/mlx/issues/1004 I saw this implementation too (without an inverse) https://github.com/nuniz/mlx_stft
— Reply to this email directly, view it on GitHub https://github.com/keras-team/keras/issues/19571#issuecomment-2599408622, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAFNM37QZKZSTXJGJWT7ADD2LGNJLAVCNFSM6AAAAABVICBTNGVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDKOJZGQYDQNRSGI . You are receiving this because you were mentioned.Message ID: @.***>
Comment From: acsweet
I'm going to hold off on math.qr
for now, mlx currently only supports square matrices (and no option for the complete or reduced factorization).
I have a PR for fft2
, rfft
, and irfft
(and a fix to fft
), if that looks good I'll start looking at the rnn
namespace.
It looked like the backend implementations for rnn.gru
and rnn.lstm
were only implemented for tensorflow for cudnn specific speedups with tf. So I think it's safe to follow similarly to jax and torch?
Comment From: fchollet
Right, unless mlx actually exposes some cudnn bindings for these
On Fri, Jan 17, 2025, 11:15 PM acsweet @.***> wrote:
I'm going to hold off on math.qr for now, mlx currently only supports square matrices (and no option for the complete or reduced factorization).
I have a PR for fft2, rfft, and irfft (and a fix to fft), if that looks good I'll start looking at the rnn namespace.
It looked like the backend implementations for rnn.gru and rnn.lstm were only implemented for tensorflow for cudnn specific speedups with tf. So I think it's safe to follow similarly to jax and torch?
— Reply to this email directly, view it on GitHub https://github.com/keras-team/keras/issues/19571#issuecomment-2599588836, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAFNM367LBQFSKY3WNIPHQ32LH5S5AVCNFSM6AAAAABVICBTNGVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDKOJZGU4DQOBTGY . You are receiving this because you were mentioned.Message ID: @.***>
Comment From: acsweet
I'm going to start working through mlx.nn
now.
I hope that's okay, but I'm going to start with conv
, and if @lkarthee or @Faisal-Alsrheed would like to jump back in, please do! Otherwise I'll keep working through the other functions.
Comment From: acsweet
@awni Would it be possible to get support for non-square matrices implemented in mlx.linalg.qr
? I didn't see an open issue for it, I can open a feature enhancement too.
Comment From: awni
Yes please open an issue about it, it should be straightforward to get it working
Comment From: acsweet
If the conv
implementation looks good, I think I'll get started on the other convolutional functions
- depthwise_conv
- separable_conv
- conv_transpose
Comment From: acsweet
I have a pull request for the remaining convolutional functions, if those look good I'll continue!
Fadi asked to work on max_pool
and avg_pool
, so I'm going to work on the remaining nn
functions that are failing tests.
Comment From: acsweet
@awni Would it be pretty straightforward to implement singular value norms in linalg::matrix_norm
?
I can open an issue for it too!
Comment From: acsweet
It's in the PR, but for reference:
- psnr
, ctc_loss
, and ctc_decode
implemented
- ctc_decode
with beam_search
is not the most efficient without a unique
function from mlx
- norm
was swapped to use jax's until mlx`s supports singular value norms (https://github.com/ml-explore/mlx/issues/1791)
I'm going to continue with numpy
and linalg
implementations focusing on passing tests in keras/src/layers
Comment From: fbadine
pooling functionality added in https://github.com/keras-team/keras/pull/20814
Comment From: acsweet
Once these latest two PRs are merged, I'd like to try merging the master branch into mlx (if that's okay)
Comment From: acsweet
Merged the Keras master
branch into mlx
and patched a few files for pytest to work
Going to add new functions and check tests starting with nn.py
Will start adjusting related tests that should be skipped for mlx
, e.g. float64
, flash_attention
, etc.
Comment From: acsweet
I'm currently working through getting the layer tests to pass (keras/src/layers
) including updates to ops
functions as needed, and skipping unsupported tests.
Updates to ops
are mostly in math
, nn
, image
, numpy
, and core
.
Comment From: acsweet
We're continuing to work through the remaining functions in linalg
and numpy
, and adjusting tests where appropriate.
I'm marking quantize tests as unsupported by mlx for now. The quantized matmul with int8
is quite strict, and float8
isn't supported yet. If anyone has some other thoughts on this, I'd be very happy to hear.
Comment From: awni
The quantized matmul with int8 is quite strict
Just curious what you mean by that? What flexibility is missing?
Comment From: acsweet
Sorry Awni, I think I spoke too soon! I'm still wrapping my head around mlx's quantization and Keras' quantization related methods (and quantization in general).
It looked like to call quantize
the columns of the input needed to be divisible by the group_size
(either 64 or 128), and quantized_matmul
was with one non-quantized array. Can matrix multiplication be performed with two quantized arrays? I think mlx.matmul
only supports floating point types, will this method allow for integer types too at some point?
I think I need to read up more and maybe pick someone's brain on this topic soon!
Comment From: awni
It looked like to call quantize the columns of the input needed to be divisible by the group_size (either 64 or 128)
Yes (32, 64 or 128 are supported)
and quantized_matmul was with one non-quantized array
Yes.
Can matrix multiplication be performed with two quantized arrays?
Not yet
I think mlx.matmul only supports floating point types, will this method allow for integer types too at some point?
I think so but it's not been a top priority.
I think you are probably right that quantization might be difficult to support across platforms. Quant formats and options are quite diverse, there isn't a standard yet.
Comment From: fbadine
New release of MLX (0.23.0) with mx.float64 support for CPU, non square QR factorisation among the introduced features.
https://github.com/ml-explore/mlx/releases/tag/v0.23.0
Comment From: fbadine
An issue was raised with MLX team regarding an issue with solve_triangular (https://github.com/ml-explore/mlx/issues/1871) A new PR is raised in MLX to tackle this issue https://github.com/ml-explore/mlx/pull/1876. When a new MLX is released that includes the fix, I will add the solve_triangular support. On the other hand solve works fine and will add it in a PR soon.
Comment From: fbadine
@awni is there any plan to support lu_factor for a non-square array?
Comment From: awni
No plan but we can do it if it’s useful. Please file an issue
Comment From: fbadine
No plan but we can do it if it’s useful. Please file an issue
This is now implemented in mlx https://github.com/ml-explore/mlx/pull/1889 Once it is released, I will add lu_factor support
Comment From: fbadine
test_argmax_neg and test_argmin_negative_zero are failing in numpy_test.py::NumpyOneInputOpsDynamicShapeTest due to a bug MLX where the values are wrong on GPU. An issue was raised with MLX team.
https://github.com/ml-explore/mlx/issues/1895
Comment From: fbadine
test_argmax_neg and test_argmin_negative_zero are failing in numpy_test.py::NumpyOneInputOpsDynamicShapeTest due to a bug MLX where the values are wrong on GPU. An issue was raised with MLX team.
https://github.com/ml-explore/mlx/issues/1895
This turned out to be normal behaviour on metal with no flag to disable it. We need to skip those tests for MLX. I will do so in my next PR.