Checking job availability...
Original
Simplified
Job Overview: We are seeking an experienced Module Lead to guide and lead a team of developers in designing, developing, and implementing solutions in Natural Language Processing (NLP), ETL processes, and Hadoop technologies. As a Module Lead, you will play a key role in driving the architecture, strategy, and execution of complex systems and data pipelines. You will manage the technical direction of the team, mentor junior developers, and collaborate with cross-functional teams to ensure successful delivery of projects. Key Responsibilities: Lead the design and development of NLP-based solutions, including text classification, named entity recognition, sentiment analysis, and other advanced NLP tasks. Oversee the development and implementation of robust ETL (Extract, Transform, Load) processes for data integration, transformation, and migration. Guide the team in building and optimizing Hadoop-based solutions, ensuring efficient data storage, processing, and analysis. Architect scalable, high-performance solutions using Hadoop and related technologies (MapReduce, Hive, Pig, HBase). Provide technical leadership and mentorship to junior and mid-level developers, ensuring adherence to best practices in coding, design, and deployment. Collaborate with data scientists, business analysts, and other stakeholders to understand business requirements and translate them into technical specifications. Ensure the quality and performance of NLP models, ETL pipelines, and Hadoop jobs by implementing automated testing and performance tuning. Manage the integration of multiple data sources and ensure seamless data flow across systems. Monitor project progress, identify risks, and ensure timely delivery of solutions. Drive innovation and improvements in NLP, ETL, and Hadoop practices across the team. Document technical designs, architecture, and processes for internal and external stakeholders. Stay current with advancements in NLP, big data technologies, and best practices. Skills and Qualifications: Proven experience in NLP technologies and frameworks (e.g., spaCy, NLTK, Transformers, BERT). Strong expertise in ETL processes, including data extraction, transformation, and loading, with experience in tools like Talend, Informatica, or Apache NiFi. Hands-on experience with Hadoop ecosystem, including MapReduce, Hive, Pig, HBase, and Spark. Experience with data warehousing, data lakes, and data integration practices. Proficient in programming languages such as Java, Python, Scala, or R. Strong problem-solving and analytical skills with a focus on optimization and scalability. Experience with version control systems (e.g., Git), continuous integration, and deployment pipelines. Familiarity with cloud platforms (e.g., AWS, Azure, Google Cloud) and Hadoop deployments on cloud services is a plus. Excellent communication and interpersonal skills to work with cross-functional teams and stakeholders. Strong leadership skills with the ability to manage and motivate a team. Ability to work in an Agile environment, handling multiple tasks and meeting deadlines. Bachelor's degree in Computer Science, Engineering, Data Science, or a related field (or equivalent work experience). Preferred Qualifications: Advanced degree (Master's or PhD) in Computer Science, Data Science, or related fields. Experience with containerization and orchestration tools (e.g., Docker, Kubernetes). Knowledge of machine learning models and frameworks (e.g., TensorFlow, PyTorch) applied to NLP. Familiarity with NoSQL databases (e.g., MongoDB, Cassandra) is an advantage.