Cognizant logo

Databricks

Cognizant
Full-time
On-site
Tamil Nadu
Technology & Engineering


Job Summary

Join our dynamic team as a Developer where you will leverage your expertise in Spark in Scala Databricks and Amazon technologies to drive innovative solutions. With a hybrid work model and day shifts you will collaborate with cross-functional teams to optimize data processes and enhance our data infrastructure. Your contributions will directly impact our companys success and societal advancements.


Responsibilities

  • Develop and implement scalable data processing solutions using Spark in Scala to enhance data efficiency and performance.
  • Collaborate with cross-functional teams to design and optimize data workflows using Databricks Workflows and Databricks SQL.
  • Administer and manage Databricks Unity Catalog to ensure data governance and security across the organization.
  • Utilize Databricks CLI to automate data pipeline deployments and streamline operational processes.
  • Integrate and manage data storage solutions using Amazon S3 and Amazon Redshift to support data analytics initiatives.
  • Write and maintain high-quality Python and PySpark code to process and analyze large datasets effectively.
  • Implement Databricks Delta Lake for reliable and efficient data storage and retrieval.
  • Monitor and troubleshoot data workflows to ensure seamless data operations and minimize downtime.
  • Provide technical expertise and support to team members fostering a collaborative and innovative work environment.
  • Stay updated with the latest industry trends and technologies to continuously improve data solutions.
  • Contribute to the development of best practices and standards for data engineering processes.
  • Ensure data quality and integrity through rigorous testing and validation procedures.
  • Document and communicate technical solutions and processes to stakeholders for transparency and alignment.


Qualifications

  • Possess strong experience in Spark in Scala and Databricks technologies demonstrating proficiency in data engineering.
  • Have hands-on experience with Amazon S3 and Amazon Redshift for data storage and management.
  • Demonstrate expertise in Python and PySpark for data processing and analysis.
  • Experience with Databricks SQL and Delta Lake is essential for efficient data operations.
  • Familiarity with Databricks CLI and Unity Catalog Admin is highly desirable.
  • Strong problem-solving skills and ability to work in a hybrid work model.