MUXU

Edge AI & IoT Applications

On-device AI inference and IoT application services for embedded Linux.

Services

From model optimization to full IoT architecture โ€” deploying AI and connected applications on embedded devices.

On-device LLM inference and deployment

Local LLMs for privacy-preserving, low-latency NLP without cloud dependency.

Model optimization and quantization

INT8/INT4 quantization, pruning, distillation for constrained devices.

Neural network accelerator integration

NPU, GPU, TPU integration for inference throughput and energy efficiency.

Sensor fusion and data preprocessing

Combining multi-sensor data into clean inputs for AI models.

Edge-to-cloud AI pipelines

Hybrid architectures: edge inference + cloud model updates and analytics.

IoT application architecture

End-to-end IoT design: protocols, data flow, device management.

Frameworks and runtimes

Open source frameworks and runtimes for deploying machine learning models and large language models on embedded devices.

Inference engines

Local inference engines for running large language models on edge hardware.