Description & Requirements
Bloomberg’s internal and enterprise compute and data science solutions were established to support development efforts around data-driven compute, machine learning, and business analytics. Both the Data Science Platform and BQuant Platform are solutions that aim to provide scalable compute, specialized hardware and first-class support for a variety of workloads such as Spark, Trino, and Jupyter. These solutions are built using containerization, container orchestration and cloud architecture.
As the needs of distributed compute, machine learning, data exploration and analysis advance, so do the needs of the compute solution that underscores it. Accentuated by the widespread success of Large-Language-Models and AI initiatives across Bloomberg, these platforms are poised for continued growth to accommodate the endless number of products across Bloomberg that rely on a robust compute environment. Highlights from our upcoming roadmap focus on creating a highly scaled and performant compute solution that abstracts away common requirements that appear across many use cases, including creating a highly available federation layer for Batch Spark Workloads, increasing compute resource usage efficiency and visibility, enhancing the Interactive Spark experience, and continuing to enhance our cloud integration within BQuant’s infrastructure.
As a member of the Spark Engineering Team, you’ll have the opportunity to make key technical decisions to keep these solutions moving forward. Our team makes extensive use of open source (e.g. Spark, Kubernetes, Istio, Calico, Buildpacks, Kubeflow, Jupyter etc.) and is deeply involved in a number of communities. We collaborate widely with the industry, contribute back to the open source projects, and even present at conferences. While working on the platform, the backbone for many of Bloomberg's up and coming products, you will have the opportunity to collaborate with engineers across the company and learn about the technology that delivers products from the news to financial instruments. If you are a software engineer who is passionate about building resilient, highly available infrastructure and seamless, usable full stack solutions, we'd like to talk to you about an opening on our team.
We’ll trust you to:
Interact with data engineers and ML experts across the company to assess their development flow and scale requirements
Solve complex problems such as cluster federation, compute resource management and public cloud integration.
Build first-class observability in a cloud-native way that provide insights that our users need
Educate users through tech talks, professional training, and documentation
Collaborate across data science teams on proper use/integration of our platform
Tinker at a low level and communicate your work at a high level
Research, architect and drive complex technical solutions, consisting of multiple technologies
Mentor junior engineers and be a strong engineering voice who takes charge driving part of Spark’s technical vision
You’ll need to have:
4+ years of programming experience with at least 2 object-oriented programming languages (Go, Python, Java) and willingness to learn more as needed
A degree in Computer Science, Engineering or similar field of study or equivalent work experience.
Experience building and scaling container-based systems using Kubernetes
Experience with distributed data analytics frameworks eg. Spark, Trino, Presto, Kafka
Ability to keep up with open source tech and trends for data analytics
A passion for providing reliable and scalable enterprise-wide infrastructure
We’d love to see:
Experience with Kubebuilder and Kubernetes operator-based frameworks
Experience working with platform security standards such as Spiffe and Spire
Experience with mainstream machine learning frameworks such as PyTorch, Tensorflow
Open source involvement such as a well-curated blog, accepted contribution, or community presence
Experience operating production systems in the public cloud e.g. AWS, GCP, or Azure
Experience with configuration management systems (e.g. Babka)
Experience with continuous integration tools and technologies (Jenkins, Git, Chat-ops)
If this sounds like you, apply! You can also learn more about our work using the links below:
Managing Multi-Cloud Apache Spark on Kubernetes https://bburl/hXNER
Scaling Spark on Kubernetes - https://bburl/e8ZnC
Kubeflow for Machine Learning: https://bburl/uJnFT
HDFS on Kubernetes: Tech deep dive on locality and security: https://bburl/Ymbcs
Apache Spark on k8s and HDFS Security: https://bburl/DD6NW
Machine Learning the Kubernetes Way - https://bburl/E7LVp
Inference with KFServing - https://bburl/jv8Fv
ML at Bloomberg - https://bburl/TuT42
Introducing KFServing - https://bburl/XaBz5
Kubernetes on Bare Metal - https://bburl/B6HXS
Serverless Inferencing on Kubernetes - https://bburl/Xvqq3
Serverless ML Inference https://bburl/bECkj
We offer one of the most comprehensive and generous benefits plans available and offer a range of total rewards that may include merit increases, incentive compensation, [Exempt roles only], paid holidays, paid time off, medical, dental, vision, short and long term disability benefits, 401(k) +match, life insurance, and various wellness programs, among others. The Company does not provide benefits directly to contingent workers/contractors and interns.