Citi's GSP/Rates businesses require advanced business analysis, strategic project management, cutting-edge application development, robust quality assurance, and seamless integration with proprietary technology solutions. The Realtime Risk Data team needs to architect, build, and operate high-performance, resilient data platforms for critical financial operations and real-time decision-making.
Requirements
- Expert-level command of Java (version 11 or higher) with a deep understanding of concurrent programming, multithreading, advanced OOP concepts, design patterns, and performance tuning.
- Proven, hands-on production experience and deep architectural understanding of: Apache Kafka or related technologies: For high-throughput, fault-tolerant message queuing and streaming.
- Proven, hands-on production experience and deep architectural understanding of: Apache Flink or related technologies: For advanced real-time stream processing, complex event processing (CEP), and stateful computations.
- Proven, hands-on production experience and deep architectural understanding of: Apache Pinot or related technologies: For ultra-low-latency OLAP queries on streaming data.
- Strong expertise in designing and implementing highly available, scalable, and resilient distributed systems.
- Extensive experience with large-scale data storage solutions (e.g., S3, HDFS) and distributed query engines (e.g., Trino/Presto, Spark SQL).
- Advanced SQL knowledge with experience in optimizing complex queries for large datasets.
Responsibilities
- Serve as the principal architect for scalable, high-performance Java-based real-time data solutions, ensuring robust design for high availability, fault tolerance, and resilience for both real-time and EOD risk processes.
- Drive the strategic implementation and optimization of distributed stream processing frameworks (Apache Kafka, Apache Flink) and real-time data storage technologies (Apache Pinot) for ultra-low-latency analytics and complex event processing.
- Lead the end-to-end design, development, and operation of real-time streaming data pipelines, integrating with large-scale object storage solutions like S3 and analytics engines such as Trino.
- Champion continuous improvement in data reliability, efficiency, and scalability. Establish and enforce best practices for code quality, performance optimization, and system resilience through hands-on leadership and thorough peer code reviews.
- Mentor and technically guide senior and lead developers.
- Drive significant contributions across all phases of the Agile software development lifecycle, from architectural vision and detailed design to implementation, deployment, monitoring, and ongoing support for critical real-time data systems.
- Collaborate strategically with business analysts, product managers, quality assurance teams, and other engineering leads to ensure the delivery of seamlessly integrated, high-impact technology solutions that align with business objectives and architectural standards.
Other
- 7+ years of progressive experience in data engineering and software development, with a significant focus on building high-performance, large-scale distributed systems.
- Demonstrated experience leading technical initiatives and teams within an Agile software development environment.
- Exceptional communication, analytical, and problem-solving skills, with the ability to articulate complex technical concepts to diverse audiences and drive consensus on architectural decisions.
- Experience working effectively within a geographically distributed, global development team.
- Advanced Degree: Strong academic record, ideally with a Master's or Ph.D. in Computer Science, Electrical Engineering, or a related technical/quantitative discipline.