About Telnyx
Telnyx is an industry leader that's not just imagining the future of global connectivityâwe're building it. From architecting and amplifying the reach of a
private, global, multi-cloud IP network
, to bringing
hyperlocal edge
technology right to your fingertips through intuitive APIs, we're shaping a new era of seamless interconnection between people, devices, and applications.
We're driven by a desire to transform and modernize what's antiquated, automate the manual, and solve real-world problems through innovative connectivity solutions. As a testament to our success, we're proud to stand as a financially stable and profitable company. Our robust profitability allows us not only to invest in pioneering technologies but also to foster an environment of continuous learning and growth for our team.
Our collective vision is a world where borderless connectivity fuels limitless innovation. By joining us, you can be part of laying the foundations for this interconnected future. We're currently seeking passionate individuals who are excited about the opportunity to contribute to an industry-shaping company while growing their own skills and careers.
The Opportunity
Our Python teams build highly available, low-latency microservices using modern, cloud-native technologies. This role sits within our broader AI organization and focuses on developing the
platforms, services, and tooling
that support AI-powered products â from inference and embeddings to APIs, data flows, and observability.
Youâll work in a fast-paced environment where reliability, performance, and usability matter just as much as innovation. While some squads work closer to inference pipelines and LLM tooling, others focus on platform foundations, troubleshooting and reliability, data systems, or observability â all critical to delivering AI in production at scale.
What Youâll Do
Build, test, and deploy scalable Python microservices for AI-powered systems, including text and voice processing
Design developer-friendly APIs and internal tools that make AI easy to use across the Telnyx product ecosystem
Contribute to AI inference, embeddings, and platform services that support real-time and low-latency workloads
Optimize performance, reliability, and observability in production environments
Collaborate closely with Product, Infrastructure, and ML teams to bring features from ideation to production
Troubleshoot, debug, and support critical systems running in production
Participate in code reviews and uphold high engineering and operational standards
Depending on the squad, your focus may lean more toward inference pipelines, platform foundations, observability, data systems, or production troubleshooting.
Who Weâre Looking For
3+ years of professional software development experience in Python
Strong grasp of async Python frameworks (e.g. FastAPI, Starlette)
Experience building and operating microservice-based systems
Comfort working with relational databases at scale
Solid understanding of distributed systems and production reliability
Familiarity with LLM frameworks and tools (e.g. PyTorch, Transformers, DeepSpeed, vLLM), or strong experience building backend or platform systems that support ML workloads
Excellent communication skills and the ability to work effectively across teams
Nice to Haves
Experience with distributed computing frameworks such as Ray (Ray Core, Serve, Train)
Exposure to cloud-native stacks (Docker, Kubernetes, Prometheus, Jaeger)
Knowledge of LLM optimization techniques (e.g. quantization, speculative decoding)
Background in Retrieval Augmented Generation (RAG) systems
Experience working with GPU stacks such as AMD ROCm
Curiosity and enthusiasm for staying up-to-date with modern AI and ML practices
How We Hire Across Our AI Teams
Our AI Chapter is made up of multiple squads, including:
AI application \& software engineering
Platform \& infrastructure foundations
Troubleshooting, reliability, and production support
Data, storage, and embeddings
Observability and performance monitoring
Skills
We hire into a
shared AI engineering pipeline
. During interviews, we focus on core engineering skills, system design, and problem-solving ability. Final squad alignment happens later, based on:
Your strengths and interests
Team needs at the time
Where youâll have the biggest long-term impact
This means the exact focus of the role may vary by squad, but the core expectation remains the same:
building reliable, scalable AI systems in production
.
Why Youâll Love It Here
At Telnyx, Weâre Not Just Building Products â Weâre Building The Future Of Intelligent Communication. Our TelnyxAI Platform Empowers Developers And Businesses To
Store and serve vector embeddings at scale
Use open-source models with our hosted inference infrastructure
Deploy multimodal AI features across text, voice, and storage seamlessly
Youâll have the autonomy to take ownership of meaningful projects, contribute ideas that shape the product roadmap, and collaborate with talented people who care deeply about impact, quality, and long-term thinking.