4th Industrial Revolution News South Africa

Subscribe

Elections 2024

Siviwe Gwarube tells us why the DA could help South Africa succeed!

Siviwe Gwarube tells us why the DA could help South Africa succeed!

sona.co.za

Advertise your job ad
    Search jobs

    Apple enters the AI race, but prioritises on-device machine learning

    Apple’s machine learning research team has introduced a new array framework called MLX that is said to be designed for efficient and flexible machine learning on Apple's home baked processors. The framework is similar to NumPy, a popular library in Python that provides support for large, multi-dimensional arrays and matrices, along with a collection of mathematical functions to operate on these arrays.
    Apple recently unveiled its M3 family of processers in the 2023 MacBook Pro models. Source: Apple Newsroom
    Apple recently unveiled its M3 family of processers in the 2023 MacBook Pro models. Source: Apple Newsroom

    The Python API of MLX is very similar to that of NumPy, with a few exceptions. Additionally, MLX also offers a fully featured C++ API that closely follows the Python API. While both MLX and NumPy are array frameworks, there are several key differences between them.

    MLX allows for the composition of function transformations for automatic differentiation, automatic vectorization, and computation graph optimization, offering a more flexible approach to mathematical operations.

    But unlike NumPy, computations in MLX are lazy, meaning they are only carried out when necessary, which can lead to efficiency improvements.

    MLX also offers multi-device compatibility, allowing operations to run on any supported devices, such as CPU and GPU, providing a more versatile platform for machine learning operations.

    This design draws inspiration from other frameworks like PyTorch, Jax, and ArrayFire. However, a notable difference between these frameworks and MLX is the unified memory model - which is a unique characteristic of Apple's ARM-based M-series SoCs.

    In MLX, arrays exist in shared memory, allowing operations on MLX arrays to be performed on any of the supported device types without the need for data copies. Currently, MLX supports device types like the CPU and GPU.

    Recent announcements

    Apple's MLX release comes in the wake of other AI announcements from its biggest competitors like Google, Microsoft and Amazon.

    Last week IBM and Amazon Web Services (AWS) announced an expansion of their partnership, aiming to help more clients operationalise and derive value from generative AI. As part of this expansion, IBM Consulting is set to deepen and expand its generative AI expertise on AWS, with a goal of training 10,000 consultants by the end of 2024.

    The two tech giants also plan to deliver joint solutions and services, upgraded with generative AI capabilities, designed to help clients across critical use cases. This move could potentially enhance the capabilities of Amazon’s AI assistant, Amazon Q, providing users with a more sophisticated and intuitive AI experience.

    A key difference between Apple's approach is that developers seem limited to how the applications work on-device and Apple isn't sending data off to its cloud servers.

    About Lindsey Schutters

    Lindsey is the editor for ICT, Construction&Engineering and Energy&Mining at Bizcommunity
    Let's do Biz