The demand for new datacenters and AI compute is rapidly outpacing the planet's energy capacity. Digital solutions are hitting a power wall as we approach the physical limits of traditional silicon. Conquering this bottleneck isn’t about bigger chips or more of them; it means rethinking the fundamental architecture. The industry's current path isn’t going to meet the need, so we took a different approach.
Instead of traditional electronic circuits, we use silicon photonics and an active, programmable metasurface to perform matrix multiplications at the speed of light. Our optical cells are 10,000x smaller than traditional photonic components, enabling unprecedented density. By using photonics instead of electricity, our chips become more efficient as they scale. This architecture will deliver up to 100 times the energy efficiency of existing solutions while significantly improving performance for large-scale AI inference.
We’ve assembled a world-class team of industry veterans and recently raised a $110M Series A led by Gates Frontier. Participants include M12 (Microsoft’s Venture Fund), Carbon Direct Capital, Aramco Ventures, Bosch Ventures, Tectonic Ventures, Space Capital, and others. We have also been recognized on the EE Times Silicon 100 list for several consecutive years.
Join us and shape the future of computing!
Position Overview:
We are seeking an experienced machine learning architect to lead the porting and optimization of large language models (LLMs), diffusion models, and other ML applications to our revolutionary optical inference engines. This role is critical to demonstrating the full potential of our metamaterial-based optical processing units (OPUs) by adapting state-of-the-art AI models to leverage our ultra-high-throughput, low-precision compute architecture. The ideal candidate will bridge the gap between cutting-edge ML research and novel hardware capabilities, ensuring customers can seamlessly deploy their AI workloads on Neurophos hardware.
Location: Austin, TX or San Mateo, CA. Full-time onsite position.
Key Responsibilities:
Lead the porting of LLM applications, diffusion models, and visual ML applications to Neurophos optical inference engines
Adapt models from diverse sources, including GitHub, Hugging Face, other open-source repositories, and customer private models
Work with models in various formats, including PyTorch, Triton, JAX, and emerging frameworks
Develop and implement quantization strategies to migrate models from higher precision formats (FP8, INT8, and above) to our optimized 4-bit precision (FP4/INT4) for weights and activations
Design and execute re-quantization, retraining, and other model adaptation techniques to minimize accuracy loss during precision reduction
Create or integrate third-party tools and workflows for efficient model porting and optimization
Optimize GEMM operations for high-throughput execution
Develop benchmarking methodologies to measure and validate model quality post-porting, including perplexity metrics and other quality indicators
Collaborate with hardware and software teams to co-optimize model architectures for optical compute characteristics
Publish research papers on novel optimization techniques and methodologies (with appropriate IP protection)
Qualifications:
MS or PhD in Computer Science, Data Science, Machine Learning, Mathematics, or related field
7+ years of experience in machine learning engineering with at least 3 years focused on model optimization and deployment
Deep expertise in neural network quantization techniques, including post-training quantization (PTQ) and quantization-aware training (QAT)
Strong proficiency in PyTorch and familiarity with other ML frameworks (JAX, Triton, TensorFlow)
Hands-on experience with transformer architectures, LLMs, and diffusion models
Experience with low-precision inference optimization (INT8, FP8, or lower)
Strong understanding of GEMM operations and linear algebra optimizations for deep learning
Experience with model evaluation metrics, including perplexity, accuracy, and benchmark suites
Track record of successfully deploying ML models on specialized hardware accelerators
Excellent communication skills with the ability to collaborate across hardware and software teams
Preferred Skills:
Experience with sub-8-bit quantization (INT4, FP4) and mixed-precision inference
Familiarity with Hugging Face Transformers library and model hub ecosystem
Experience with ONNX, TensorRT, or other model optimization frameworks
Background in analog or optical computing architectures
Knowledge of in-memory computing paradigms and matrix-vector multiplication acceleration
Published research in model compression, quantization, or efficient inference
Experience with large-scale batch inference optimization
Familiarity with prefill vs. decode optimization strategies in LLM inference
This is an opportunity to play a pivotal role in an innovative startup redefining the future of AI hardware. Work on a game-changing technology at the intersection of photonics and AI as part of a collaborative and brilliant team. You’ll contribute to a platform that redefines computational performance and accelerates the future of artificial intelligence. Come help us bring this transformative technology to the world.
BenefitsJoin a team that invests in your future and your well-being. At Neurophos, we offer:
100% coverage of base health plan premiums for you and your dependents, plus HSA contributions.
Unlimited PTO. No rigid vacation banks, just a focus on delivery.
401(k) matching and stock option opportunities to ensure our success is your success.
Full suite of voluntary benefits, including Dental, Vision, Life, Hospital, Critical Illness, and Accident insurance.
Personalized Benefits. Choose the plans that fit your life and take the cash back for those that don’t.
Top Skills
Neurophos Austin, Texas, USA Office
Austin, TX, United States
Similar Jobs
What you need to know about the Austin Tech Scene
Key Facts About Austin Tech
- Number of Tech Workers: 180,500; 13.7% of overall workforce (2024 CompTIA survey)
- Major Tech Employers: Dell, IBM, AMD, Apple, Alphabet
- Key Industries: Artificial intelligence, hardware, cloud computing, software, healthtech
- Funding Landscape: $4.5 billion in VC funding in 2024 (Pitchbook)
- Notable Investors: Live Oak Ventures, Austin Ventures, Hinge Capital, Gigafund, KdT Ventures, Next Coast Ventures, Silverton Partners
- Research Centers and Universities: University of Texas, Southwestern University, Texas State University, Center for Complex Quantum Systems, Oden Institute for Computational Engineering and Sciences, Texas Advanced Computing Center

