Copyright OPTnation. All rights reserved.

Data Engineer

Job ID : 39263

Job Title : Data Engineer

Location : New York, NY

Comapny Name : Apollose

Job Type : Full-Time, Part-Time, Contract, Training

Industry : Information Technology

Salary :  $50000 - $90000  per year

Work Authorization : ["OPT","CPT","Entry Level","F1","H4","L1","H1 Visa","TN Permit Holder","Green Card Holder","Canadian Citizen","All EAD","US Citizen"]

No. of Positions : I have ongoing need to fill this role

Posted on : 07-10-2025

Required Skills : Java, C++, or Python

Benefits : Medical Insurance, Dental Insurance, Vision Insurance, 401K, Life Insurance

Job Description :

KEY RESPONSIBILITIES:

  • Our Big Data capability team needs hands-on developers who can produce beautiful functional code to solve complex analytics problems. If you are an exceptional developer with an aptitude to learn and implement using new technologies, and who loves to push the boundaries to solve complex business problems innovatively, then we would like to talk with you.
  • You would be responsible for evaluating, developing, maintaining, and testing big data solutions for advanced analytics projects.
  • The role would involve big data pre-processing reporting workflows including collecting, parsing, managing, analyzing and visualizing large sets of data to turn information into business insights.
  • The role would also involve testing various machine learning models on Big Data and deploying learned models for ongoing scoring and prediction. An appreciation of the mechanics of complex machine learning algorithms would be a strong advantage.

SKILLS & QUALIFICATIONS:

  • Demonstrable experience designing technological solutions to complex data problems, developing testing modular, reusable, efficient and scalable code to implement those solutions.
  • Ideally, this would include work on the following technologies:
  • Expert-level proficiency in at-least one of Java, C++, or Python (preferred). Scala knowledge a strong advantage
  • Strong understanding and experience in distributed computing frameworks, particularly Apache Hadoop 2.0 (YARN; MR HDFS) and associated technologies -- one or more of Hive, Sqoop, Avro, Flume, Oozie, Zookeeper, etc.
  • Hands-on experience with Apache Spark and its components (Streaming, SQL, MLLib) is a strong advantage.
  • Operating knowledge of cloud computing platforms (AWS, especially EMR, EC2, S3, SWF services and the AWS CLI)
  • Experience working within a Linux computing environment, and use of command line tools including knowledge of shell/Python scripting for automating common tasks.
  • Ability to work in a team in an agile setting, familiarity with JIRA and clear understanding of how Git works.
  • Linux environment and shell scripting.

Company Details :

Company Information hidden please Login to view details

Login To Apply Now! Register & Apply Now!