Job Description
Job Responsibilities: • Install, configure, and deploy the Databricks Unified Analytics Platform in accordance with best practices and organizational requirements. Set up clusters, workspaces, notebooks, and other system objects as needed. • Manage user accounts, groups, permissions, and access controls within the Databricks platform. Define role-based access controls (RBAC). • Configure and manage Databricks clusters to optimize performance, scalability, and resource utilization. Monitor cluster health, adjust cluster configurations, and scale clusters up or down as needed to meet workload demands. • Monitor and optimize the performance of Databricks workloads, queries, and jobs. Identify and address performance bottlenecks, optimize Spark configurations, and fine-tune resource allocation for optimal performance. • Implement and enforce security measures to protect data and resources within the Databricks platform. • Provide support on network security, encryption, authentication, and auditing controls to ensure compliance with regulatory requirements and security policies. • Integrate Databricks with other data sources, data warehouses, and analytics tools as part of the data ecosystem. Automate routine tasks and workflows using Databricks APIs, command-line interfaces (CLIs), and scripting languages. • Monitor system health, resource utilization, and performance metrics using monitoring tools and dashboards. Set up alerts and notifications. • Develop and implement backup and recovery strategies to protect against data loss and ensure business continuity. Perform regular backups, validate backups, and test recovery procedures to minimize downtime. • Forecast future growth and resource requirements based on current usage patterns and business projections. Scale Databricks resources such as clusters and storage as necessary.
Responsibilities
Required Experience & Skills: • Specific experience in administering Databricks Unified Analytics Platform, minimum of 4 years of experience. • Strong understanding of big data technologies, distributed computing, and Apache Spark architecture. • Hands-on experience with cloud computing platforms such as AWS, Azure, or Google Cloud Platform. • Familiarity with infrastructure as code (IaC) tools such as Terraform or CloudFormation. • Excellent troubleshooting and problem-solving skills, with the ability to diagnose and resolve complex technical issues. Strong communication and collaboration skills, with the ability to work effectively in cross-functional teams.

