d-Matrix is looking for a AI System Solutions Architect to develop world-class products around d-Matrix inference accelerators. In this role you will be engaged with key customers and internal architects, and other key internal and external stakeholders to drive overall System Solutions. This requires technically analyzing, defining outside-in usage cases, and use broad spectrum of technologies to drive a AI server system solution spanning silicon, platform HW/SW, and usages to deliver the best customer experiences with d-Matrix inference accelerators.
Responsibilities
- Design, develop, and deploy scalable GenAI inference solutions with d-Matrix accelerators
- Work closely with team members across architecture, engineering, product management and business developments to optimize the d-Matrix System Solutions for best performance & power balance, feature set and overall system cost.
- Work closely with Datacenter, OEM and ODM customers at early stage of product concept and planning phase, to enable the system design with partners and industrial ecosystem.
- Influence and shape the future generations of products and solutions by contributing to the system architecture and technology through the early engagement cycle with customers and industrial partners.
- Stay abreast of the latest advancements in GenAI hardware and software technologies and assess their suitability for integration into d-Matrix GenAI inference solutions.
- Establish credibility with both engineering and leadership counterparts at top technology companies, communicate technical results and positions clearly and accurately, and drive alignment on solutions.
Experience
- 5+ years of AI Server System experience by working on multiple projects from architecture, development, design including memory, I/O, power delivery, power management, boot process, FW and BMC/hardware management through bring-up and validation and supported through the release to production.
- 5+ years of experience in a customer-facing role interfacing with OEMs, ODMs and CSPs.
- Detailed understanding of server industry standard busses, such as DDR, PCIe, CXL and other high-speed IO protocol is required.
- Ability to work seamlessly across engineering disciplines and geographies to deliver excellent results.
- Deep understanding of datacenter AI infrastructure requirements and challenge
Preferred Experience
- Hands-on understanding of AI/ML infrastructure and hardware accelerators
- Experience with leading AI/ML frameworks such as PyTorch, TensorFlow, ONNX, etc. and container orchestration platforms such as Kubernetes
- Outstanding communication and presentation skills
Education
- 15 + Years of Industry Experience and Engineering degree in Electrical Engineering, Computer Engineering, or Computer Science with extensive experience.