Data Engineer Lead/architect

Year    Bangalore, Karnataka, India

Job Description


OverviewData engineer Lead/architect:OVERALL 14+ years HANDS-ON EXPERIENCE ON Big DataPayment gateway experienceResponsibilities

  • Define and achieve the strategy roadmap for Big Data architecture and design and overall enterprise data governance and management across the enterprise and Data Centers.
  • Establish standards and guidelines for the design & development, tuning, deployment and maintenance of information, advanced data analytics, ML/DL, data access frameworks and physical data persistence technologies.
  • Research in the areas of advanced data techniques, including data ingestion, data processing, data integration, data access, data visualization, text mining, data discovery, statistical methods, database design and implementation.
  • Experience in developing use cases, functional specs, design specs, ERDs etc.
  • Experience with the design and development of multiple object-oriented systems.
  • Experience with extending Free and Open-Source Software (FOSS) or COTS products.
  • Architect solutions for key business initiatives ensuring alignment with future state analytics architecture vision.
  • Work closely with the project teams as outlined in the SDLC engagement model to provide guidance in implementing solutions at various stages of projects.
  • Engage constructively with project teams to support project objectives through the application of sound architectural principles.
  • Develop and validate that the proposed solution architecture supports the stated & implied business requirements of the project.
  • Review technical team deliverables for compliance with architecture standards and guidelines.
  • Adopt innovative architectural approaches to leverage in-house data integration capabilities consistent with architectural goals of the enterprise.
  • Create architectural designs for different stakeholders that provide a conceptual definition of the information processing needs for the delivery project.
  • Provide right-fit solution to implement variety of data, analytics requirements using Big Data technologies in Insurance/Healthcare space.
  • Provide expertise in defining Big Data Technical Architecture and design using Hadoop, NoSQL and Visualization tools/platforms.
  • Ability to get down to the programming/code level and provide hands-on expertise in technical features of Big Data tools/platforms.
  • Lead a team of designers/developers and guide them throughout the system implementation life cycle.
  • Engage client Architects, Business SMEs and other stakeholders during Architecture, Design and implementation phases.
  • Define new data infrastructure platform to capture vehicle data. Research and develop new data management solutions, approaches and techniques for data privacy and data security, and advanced Data Analytics systems.
  • Design complex and distributed software modules using Big Data technologies, Streaming Data Technologies and Java/JEE for real-time stream/event data processing and real-time parametrized and full-text searching of data
  • Design NoSQL data models to achieve strong data consistency despite lack of ACID transactions
  • Design and enhance highly scalable, high performance and fault tolerant architectures across all tiers of the software and develop modules based on the architecture
  • Design and enhance the architecture that supports zero downtime requirements while different components of the system are updated/upgraded
  • Integrate with IoT devices using a data ingestion pipeline that allows application of configurable real-time rules
  • Use machine learning algorithms on the real-time data for predictive analytics
  • Design software, write code, write unit test cases, test code and review code on a daily-basis.
  • Lead multiple project modules and development teams simultaneously and ensure the quality of their deliverables and their timely delivery
  • Create/enhance scalable, high performance and fault-tolerant architectures
  • Use caching, queuing, sharding, concurrency control, etc. to improve performance and scalability
  • Enhance zero-downtime architecture during product deployment upgrades
  • Integrate with IoT devices using a data ingestion pipeline that allows application of configurable real-time rules
  • Identify the performance and scalability bottlenecks and provide solutions to resolve them
  • Perform code reviews, provide feedback and oversee code corrections to ensure compliance with the development guidelines on a daily-basis
  • Provide technical expertise in the diagnosis and resolution of issues, including the determination and provision of workaround solutions
  • Experience in data modeling and data processing architectures using NoSQL databases to achieve strong data consistency despite lack of ACID transactions
QualificationsPreferably BE/ME/BTech/MTech/MSc in computer science or related engineering field or MCAShould have relevant industry certifications like Big Data ArchitectMust possess Analytical, Interpersonal, Leadership & Organizational skills.Essential skillsRequired Skills:
  • Experience with distributed scalable Big Data store or NoSQL, including Cassandra, Cosmos DB, Cloudbase, HBase, or Big Table.
  • Experience in with Apache Hadoop and Hadoop Distributed File System (HDFS), S3 protocol, ObjectStore and with processing large data stores.
  • Experience on Hadoop, MapReduce, YARN, Zookeeper, Pig, Hive, Oozie, Flume, Sqoop, Spark and Storm, Hive, HBase, MongoDB, Cassandra
  • Experience on R, Qlik, Tableau, Postgres, Kerberos, Knox and Ranger
  • Experience on J2EE technologies including Webservices development and Enterprise Application Integration space.
  • Experience with Message Queue based technologies like RabbitMQ or ActiveMQ etc. would be considered plus point
  • Experience with other NoSQL Databases (CouchDB, Neo4j, InfiniteGraph, JanusGraph, ArangoDB, OrientDB, CockroachDB, Redis, Vitess etc.)
  • Real-time stream data processing frameworks: Beam, Spark Streaming, Flink, Kafka Stream, Apex, Heron, Storm etc.
  • Experience in Queuing: Kafka, ActiveMQ, RabbitMQ etc.
  • Experience with ETL design using tools Informatica, Talend, Oracle Data Integrator (ODI), Dell Boomi or equivalent
  • Experience with Big Data & Analytics solutions Hadoop, Pig, Hive, Spark, Spark SQL Storm, AWS (EMR, Redshift, S3, etc.)/Azure (HDInsight, Data Lake Design, Analytical services/Power BI)
  • Experience in developing code around Hadoop with Oozy, Sqoop, Pig, Hive, HBase, Avro, Parquet, Spark, NiFi
  • Experience with NoSQL platform (eg, Spark, Cassandra, Hadoop).
  • Experience with Lucene based cluster technology (eg, Elastic Search, Solr).
  • Experience with One of the queueing system (eg, Apache Kafka, RabbitMQ).
  • Experience with Cloud Computing / Distributed Computing / Amazon Web Services / GCP / AZURE
  • Experience with RESTful Web Services (eg, Jetty) & Java Spring framework
ExperienceOVERALL 14+ years HANDS-ON EXPERIENCE ON Big Data

Expleo

Beware of fraud agents! do not pay money to get a job

MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.


Related Jobs

Job Detail

  • Job Id
    JD3339228
  • Industry
    Not mentioned
  • Total Positions
    1
  • Job Type:
    Full Time
  • Salary:
    Not mentioned
  • Employment Status
    Permanent
  • Job Location
    Bangalore, Karnataka, India
  • Education
    Not mentioned
  • Experience
    Year