Mapjects.com logo

Big Data Analyst e.g. EC2, S3, AWS

Mapjects.com
On-site
Washington, District of Columbia, United States

Company Description

Mapjects is a leading centralized logistics operations portal platform. The platform servers franchises with ERP components that suite the franchise business needs. Mapjects Clearview platform provide one-click distribution, logistics and analysis products to enrich and visualize big data sets from warehousing, fulfillment, fraud detection, payment technology and b2b eCommerce.

 

Email WORD resume and contact to dev@mapjects.com or [email protected] 

  • Green Card and US citizens only
  • Compensation is very competitive, and commensurate with experience, loc DC or VA
  • Corp 2 Corp, 1099, or W2 (please email your expectations, and availability to start)
  • After face to face, and a certain period, the option to work remotely and telecommute is a possibility.
  • When can you start/availability

Job Description

Participate in a team to design and implement large scale data architecture to support evolving analysis in a variety of aviation domains. This architecture will support large scale batch processing, streaming data capture with applications to real time analytics, and back-ends to support executive views and interactive queries. Responsibilities include developing processes to rapidly fuse new data streams, integrate new metrics, and facilitate discovery of novel insights. In addition, the person fulfilling this role must understand the needs and objectives of the data consumers and develop tools to facilitate data driven insights. The successful applicant will also be responsible for performing cost benefit analysis when choosing and assessing technology alternatives and must keep abreast of developments in the open source and commercial data analytics communities.

Qualifications

Required Skills: Expertise applying some subset of the following technologies:

• languages - java, python, clojure, scala
• hands on experience with hadoop and map/reduce
• data serialization - protocol buffers, avro, thrift, trevni, hdf5, netcdf
• workflow - pig, oozie, cascading, cascalog, crunch, sqoop
• databases - HBase, Accumulo, Mongo, Cassandra, Neo4j, Impala
• business integration/analytics - splunk, platfora, pentaho, tableau, informatica, spotfire
• familiar with open source tools and development and comfortable working in Linux environment
• cloud - aws s3, ec2, elastic map/reduce 

• data modeling
• develop processes for automated data capture, processing, transformation and fusion
• design and implement data architecture to support interactive, real time and batch processing
• ensure data consistency, discoverability, and quality

Additional Information

please email word copy of resume to

ensure it has your contact information, and phone number to reach you.

[email protected] or [email protected]