POSITION TITLE: Big Data Architect
BUSINESS/ORGANIZATION: IT Corporate Applications
JOB FAMILY: Business Intelligence & Database

This position, as a member of the Business Intelligence & Database team, will play a key role in the Design and Development of our Big Data Analytics Platform and be part of the global team to support the platform. The Data Architect is an accomplished technical leader, proactive customer-focused advocate, a team player with substantial software engineering experience, preferably with experience in the apparel/fashion/retail industry. The Data Architect must have hands-on experience with enterprise level software development, integration and implementation with big data. The ideal candidate will have an advanced understanding of Java, data ingestion, data discovery & analysis, data cleansing, data transformation, data visualization, and SQL/data modelling. The Big Data Architect will work in close collaboration and cooperation with the entirety of the organization, especially the BI & Analytics team in North America
  • Design and build high performance, scalable solutions for large volume data ingestion, clean-up and processing that meets the needs of different business use cases
  • Design large-scale data processing systems, is an expert in data warehousing solutions and should be able to work with the latest database technologies, including NoSQL databases
  • Embrace the challenge of dealing with petabytes of data on a daily basis
  • Be a self-organized member of an engineering team operating in an agile environment – practice and enforce agile and scrum methodologies
  • Collaborate with cross-functional teams including product managers and business stakeholders to develop and rollout new features/capabilities to big data and analytics platform
  • Stay abreast with latest developments in big data, analytics and cloud technology areas, and bring new ideas to the table
  • Design and Develop prototypes and proof of concepts on various technologies and make recommendations on tools/platform choices
  • Design deployment automation and operational support strategies
  • Participate in the review of code and/or systems for proper design standards, quality and functionality
  • Responsible for technical documentation in accordance with department standards and industry best practices
  • Design data and metadata policies and procedures 

  • 10+ years of systems/application design and development experience
  • Must have implemented at least 2-3 assignment with Informatica BDM and the Project should have gone Live to Production
  • Strong experience in delivering at-least 3 or more assignment in Big Data related projects using Spark, Scala, Python, Hive, Hbase, Kafka, etc.
  • Must have supported at least 3 or More assignment as Solution Lead in designing large scale data lake solutions
  • Must know Data Injection Framework and Data Lake (Big Data Technologies) 
  • Strong foundation in building enterprise scale applications using APIs, batch jobs, parallel and distribution processing models
  • Good knowledge and hand-on experience in the past in delivering at-least 3 traditional DWBI, ODS & client 360 solutions using Informatica (ETL), Oracle & scheduling tools like Tivoli/Cloud Scheduler
  • 5+ years of experience in designing, and building solutions using a big data distribution, preferably MapR , for 
    • data ingestion, cleansing,Masking,Profiling and transformation (e.g. Informatica BDM,BDQ,TDM)
    • data discovery & analysis using querying tools (e.g. Impala, Hive) 
    • data storage using distributed databases (MapR, HBASE, Kudu) 
    • data streaming (e.g. Kafka, Apache Spark) 
    • data visualization (e.g. Tableau, Qlik, Lumira) 
    • processing monitoring (e.g. MapR manager, Hue)
    • Optional Informatica Stack :-- IIR, PIM,Vibe, BRDM, Real Time edition
  • 3+ years of experience building applications in at least one of the three major cloud platforms – GCP, AWS or Azure (GCP preferred)
  • Experience with NoSQL database technologies like MongoDB or Cassandra
  • Proficiency in Big Data Virtualization and Modelling tools like SAP VORA
  • Experience with object-oriented design and development patterns as well as experience in engineering (commercial or open source) software platforms and large-scale data infrastructures
  • Experience building systems that consume, organize and analyze JSON and XML messages as data.
  • Flair for data architecture and data modelling, and strives to bring efficiency in big data and analytics life cycle
  • Capable of architecting and building highly scalable distributed systems, using different open source tools 
  • Dynamic team player with the ability to collaborate with in and across teams and departments
  • Ability to iterate quickly in an agile environment
  • Ability to learn new technologies and evaluate multiple technologies to solve a problem
  • Strong quality focus; ability to refactor early and often to continuously improve code quality
  • Excellent analytical and troubleshooting skills
  • Good written and verbal communication skills in English
  • Strong work ethic and entrepreneurial spirit

Preferred Job Skills
  • Proficiency with relational technologies that supplement RDBMS tool sets
  • Knowledge in Python and/or R scripting
  • Experience in real-time analytics applications
  • Experience handling customer and marketing analytics related data sets
  • Experience in advanced analytics including predictive analytics, artificial intelligence, machine leaning etc. a big Plus
  • Experience integrating big data platforms with marketing automation platforms (e.g. campaign management software)

  • 0

    1. 0 ratings
    2. 0 reviews
  • 40% Complete (success)
  • 40% Complete (success)
  • 40% Complete (success)
  • 40% Complete (success)
  • 40% Complete (success)
Write Review