In this role you will work in a small, dynamic team with some of New York City’s top engineers, to solve very unique, large scale challenges.
You will be joining a team build a development toolset for teams making use of a central reference data infrastructure, providing core data and analytics solutions to widely distributed environments.
Leverage C++ and Java to design and develop low latency data systems
Use Hadoop and Spark for data processing and near real-time analytics solutions
Utilize Hadoop-based storage solutions
Build out new tools to manage server farms and ensure availability and efficient load balancing
Enhance infrastructure for low latency and high throughput data retrieval
3+ years experience of C++ or Java on UNIX/Linux environment
Solid understanding of algorithms and data structures
Experience with high volume, high availability distributed systems
Experience working in a test-driven development and agile environment
Knowledge of low-level Linux, UNIX and C system high performance design
Experience with systems-level Scala and Java solutions
Experience with Hadoop and using distributed analytics such as Apache Spark