The NetsPresso Platform Team at Nota AI designs and implements core platforms and software that transform advanced AI model lightweighting and optimization research into real-world products.
Our organization consists of Model Representation, Quantization, Graph Optimization, Model Engineering, and SW Engineering units. Specifically, the Model Representation Part unifies models from various deep learning frameworks into our proprietary NPIR (NetsPresso Intermediate Representation). This foundational step enables the application of optimization technologies such as quantization, graph optimization, and compression.
We secure technical versatility by integrating diverse framework models into NPIR, scale our support for the latest AI models and hardware, and maximize inference efficiency on actual hardware by optimizing the logical structure of deep learning models for target runtime environments.
As a member of this position, you will participate in the productization of the unified IR that serves as the foundation of NetsPresso. You will contribute to designing and developing NPIR by analyzing the characteristics of various deep learning frameworks and hardware.
You will gain hands-on experience applying on-device AI lightweighting and optimization techniques to actual products and solve technical challenges to ensure the latest models run efficiently on diverse hardware. As an AI Compiler Engineer, you will experience the professional fulfillment of implementing theoretical research into a robust software stack.
IR Implementation & Maintenance
Hardware-Friendly IR Transformation & Lowering
(Additional assignments may be included during the process.)
We value a strong interest in new compiler technologies and hardware backend optimization, along with the execution power to implement these into a sophisticated software stack. This role is not just about theoretical design; it is about productizing the "Lowering" and optimization process so that NPIR achieves peak performance on various hardware backends (NPU, GPU, etc.).
We emphasize a proactive collaborative attitude, communicating closely with other teams based on an understanding of hardware architecture. If you enjoy diving deep into complex operation scheduling and memory optimization to break through the technical limits of On-device AI at the hardware level, you will find great growth opportunities in our team.
The NetsPresso Platform Team at Nota AI designs and implements core platforms and software that transform advanced AI model lightweighting and optimization research into real-world products.
Our organization consists of Model Representation, Quantization, Graph Optimization, Model Engineering, and SW Engineering units. Specifically, the Model Representation Part unifies models from various deep learning frameworks into our proprietary NPIR (NetsPresso Intermediate Representation). This foundational step enables the application of optimization technologies such as quantization, graph optimization, and compression.
We secure technical versatility by integrating diverse framework models into NPIR, scale our support for the latest AI models and hardware, and maximize inference efficiency on actual hardware by optimizing the logical structure of deep learning models for target runtime environments.
As a member of this position, you will participate in the productization of the unified IR that serves as the foundation of NetsPresso. You will contribute to designing and developing NPIR by analyzing the characteristics of various deep learning frameworks and hardware.
You will gain hands-on experience applying on-device AI lightweighting and optimization techniques to actual products and solve technical challenges to ensure the latest models run efficiently on diverse hardware. As an AI Compiler Engineer, you will experience the professional fulfillment of implementing theoretical research into a robust software stack.
IR Implementation & Maintenance
Hardware-Friendly IR Transformation & Lowering
(Additional assignments may be included during the process.)
We value a strong interest in new compiler technologies and hardware backend optimization, along with the execution power to implement these into a sophisticated software stack. This role is not just about theoretical design; it is about productizing the "Lowering" and optimization process so that NPIR achieves peak performance on various hardware backends (NPU, GPU, etc.).
We emphasize a proactive collaborative attitude, communicating closely with other teams based on an understanding of hardware architecture. If you enjoy diving deep into complex operation scheduling and memory optimization to break through the technical limits of On-device AI at the hardware level, you will find great growth opportunities in our team.