Etched is building AI chips that are hard-coded for individual model architectures. Our first product (Sohu) only supports transformers, but has an order of magnitude more throughput and lower latency than a B200. With Etched ASICs, you can build products that would be impossible with GPUs, like real-time video generation models and extremely deep chain-of-thought reasoning. Software, LLM Compilation Software sells chips. Etched ASICs are no exception. While our first chip, Sohu, is only able to run transformer models, we still need production-grade software to map existing LLMs onto our chip. You will help make this happen. You will write optimized kernels for the operations that make up a transformer, like attention, model parallelism, and normalization, and package them into components that developers can use (e.g. in the way that vLLM has its fused `MergedColumnParallelLinear` component). You will work with the hardware team to debug issues that hurt performance. You will work with the software team to build integrations with existing libraries like vLLM and HuggingFace Transformers, so that our software can be drop-in compatible. You will not build a Pytorch compiler stack - instead, we will build a few highly-optimized fused kernels that can be used to implement transformer models. Representative projects: Write an optimized kernel to compute a new attention variant on our hardware Implement HuggingFace’s `CohereForCausalLM` class using Etched’s transformer building blocks Implement a synchronization mechanism to coordinate between the host CPU and Etched accelerator Implement FP8 quantization for FP16 models using the same mechanism as TransformerEngine
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Education Level
No Education Listed