Applied AI Engineer & Researcher

JigsawStack

JigsawStack

Software Engineering, Data Science
Posted on Feb 4, 2025
🧠

Applied AI Engineer & Researcher

Location: San Francisco, London, USA, Singapore

About JigsawStack:

JigsawStack provides a suite of small models that automate the boring and complex tasks in every tech stack, powering thousands of developers with millions of API requests.
At JigsawStack, our mission is to enable developers to ship better products with state-of-the-art AI models that are fast, easy to use, and accessible!

About this role:

We’re hiring our first AI Engineer & Researcher who is looking to push the boundaries of our models by implementing and experimenting with new research and techniques. As part of the JigsawStack Engineering team, you will help improve the performance and capabilities of our models by managing our AI lifecycle, including fine-tuning, automating data collection and cleaning, benchmarking, and deployment.
As a founding member at JigsawStack, your role will be dynamic, offering opportunities to spearhead and launch innovative research, papers, and products into production while being the go-to expert for all things AI!

Our tech stack:

GPU infra & frameworks: Docker, Python, PyTorch, Transformers (AWS, GCP, Azure, HuggingFace, and more)
E.g base models: Llama 3.3, YoloOCR, Llava
Frontend: NextJS/React
Backend: NodeJS
Language: Typescript
Infra: Serverless (Lambda) or Fly.io (Instances)
Database: Postgres (Supabase)
Query: SQL/GraphQL
Payment/Billing: Stripe

What you will do:

Research and understand different machine learning techniques and papers, then experiment with real-world implementation
Experience with deep learning frameworks (e.g., TensorFlow, PyTorch), and AI development tools is a must
Working alongside our engineering team to deploy models at scale on popular cloud providers like AWS, GCP, and more
Write papers on your research and experiments contributing to the growing open-source AI world
Writing detailed benchmarks on comparisons and analysis (e.g. comparing performance for two embedding models or benchmarking whisper 3 with another model)
Quantization and finetuning of LLMs and other models for our developer use cases
Optimizing models and GPU infra for cost-to-scale ratio based on load
Writing Python code and Jupyter Notebooks

Who you are:

A strong AI Researcher with at least 5 years of experience
Worked with most of our GPU infra and have a good understanding of our tech stack
Comfortable with working across the stack - model deployment, benchmarking, fine-tuning, monitoring, and scaling the service
Up-to-date with the latest papers, techniques, and experiments
Using AI to 100x your productivity
Natural curiosity to experiment with new concepts and tools while working closely with our community of developers
You care about building great products and love to geek out about the latest tech in AI
Open to feedback and have a growth mindset

Bonus if you:

Closer to the United States timezone as it’s our primary focus at our current stage
Enjoy working in a startup environment where you can take initiative and contribute to the zero-to-one phase of development
Previously ran a startup or worked as a founding team member in a startup

Benefits:

Great compensation package and equity/shares
Flexible and remote-friendly
Annually run off-sites
Learn and Grow - we provide mentorship and send you to events that help you build your network and skills
Annual Education Allowance
Let’s build benefits together

Process:

The entire process is fully remote and all communication will happen over email or via video chat.
Once you've submitted your application, the team will review your submission and may reach out for a short screening interview over a video call.
If you pass the screen you may be invited to complete a short task or discuss your past work.
If you pass the task/screen you will be invited to up to 2 follow-up interviews. The calls:
Apply for this role ->