Design and implement Big Data analytic solutions on a Hadoop-based platform. Create custom analytic and data mining algorithms to help extract knowledge and meaning from vast stores of data. Refine a data processing pipeline focused on unstructured and semi-structured data refinement. Support quick turn and rapid implementation and larger scale and longer duration analytic capability implementation.
Basic Qualifications:
* 8 years of experience with development and object-oriented programming for secure Web-based applications * 2+ years of experience with supporting a Cloud-based environment * Secret clearance * BA or BS degree
Additional Qualifications:
* Experience with Apache Solr, Spark, or Hadoop * Experience with R or Python * Experience with using repository management solutions * Experience with deploying applications in a Cloud environment * Experience with designing and developing automated analytic software, techniques, and algorithms * Experience with distributed scalable Big Data store or NoSQL, including Accumulo, Cloudbase, HBase, or Big Table * Experience in MapReduce programming with Apache Hadoop and Hadoop Distributed File System (HDFS) and processing large data stores * Experience with extending free and open-source software (FOSS) or COTS products * Ability to show flexibility, initiative, and innovation when dealing with ambiguous and fast-paced situations
Clearance:
Applicants selected will be subject to a security investigation and may need to meet eligibility requirements for access to classified information; Secret clearance is required.
We're an EOE that empowers our people-no matter their race, color, religion, sex, gender identity, sexual orientation, national origin, disability, veteran status, or other protected characteristic-to fearlessly drive change.
Let your dream job find you.
Sign up to start matching with top companies. It’s fast and free.