The On-site Azure Data Analytics Engineer at EDI Staffing will develop and optimize cloud-based Business Intelligence solutions, focusing on data integration and analytics. This role requires deep technical expertise in Azure technologies and collaboration with cross-functional teams.
ROLE SUMMARY The Azure Data Analytics Engineer will be the AZURE SME tasked with the development and optimization of cloud-based Business Intelligence solutions. Advances data analytics capabilities and drives innovative solutions. Possesses deep technical expertise in data engineering and plays instrumental role in managing data integrations from on-premises Oracle systems, Cloud CRM (Dynamics), and telematics. Collaborates closely with Data Science and Enterprise Data Warehouse teams and business stakeholders. This is an on-site position based in NJ. PRIMARY RESPONSIBILITIES: Data Ingestion and Storage: • Designs, develops, and maintains scalable, efficient data pipelines using Data Factory, and Databricks, leveraging Py Spark for complex data transformations and large-scale processing. • Builds and manages extract, transform, and load (ETL)/extract, load, transform (ELT) processes to seamlessly extract, transform, and load data from on-premises Oracle systems, customer relationship management (CRM) technology, and connected vehicles into data storage solutions, such as Azure Data Lake Storage and Azure SQL Database. Data Engineering: • Creates high-code data engineering solutions using Databricks to clean, transform, and prepare data for in-depth analysis. • Develops and manages data models, schemas, and data warehouses, utilizing Lakehouse Architecture to enhance advanced analytics and business intelligence. • Leverages Unity Catalog to ensure unified data governance and management across the enterprise's data assets. • Optimizes data storage, retrieval strategies, and query performance to drive scalability and efficiency in all data operations. Data Integration: • Integrate and harmonize data from diverse sources including on-premises databases, cloud services, APIs, and connected vehicle telematics. • Ensure consistent data quality, accuracy, and reliability across all integrated data sources. GitHub Development: • Utilizes GitHub for version control and collaborative development, implementing best practices for code management, testing, and deployment. • Develops workflows for continuous integration (CI) and continuous deployment (CD), ensuring efficient delivery and maintenance of data solutions. ADDITIONAL RESPONSIBILITIES: • Work closely with Data Science, Enterprise Data Warehouse, and Data Visualization teams, as well as business stakeholders, to understand data requirements and deliver innovative solutions. • Collaborate with cross-functional teams to troubleshoot and resolve data infrastructure issues, identifying and addressing performance bottlenecks. • Provide technical leadership, mentorship, and guidance to junior data engineers, promoting a culture of continuous improvement and innovation. REQUIRED SKILLS AND PERSONAL QUALIFICATIONS: • Technical Expertise: Extensive experience with Azure Data Factory, Databricks, and Azure Synapse, as well as proficiency in Python and PySpark. • Data Integration: Experience integrating data from on-premises Oracle systems and connected vehicle data into cloud-based solutions. • Lakehouse Architecture & Governance: Deep knowledge of Lakehouse Architecture and Unity Catalog for enterprise data governance. • Version Control & Collaboration: Demonstrated proficiency in GitHub for development, collaboration, and deployment in large-scale environments. • Infrastructure as Code (IaC): Experience with Infrastructure as Code tools such as Resource Manager (ARM) templates or terraform. • Problem-Solving & Troubleshooting: Strong analytical skills with the ability to diagnose and resolve complex data infrastructure challenges. • Collaboration: Proven ability to work effectively with Data Science teams, business stakeholders, and cross-functional teams to drive data-driven insights. • Communication: Excellent verbal and written communication skills with the ability to translate technical concepts to non-technical stakeholders. Education/Experience Requirements: BA/BS with 4 to 6 years of relevant experience. Relevant experience accepted in lieu of a degree. Work Environment • Hybrid Role: Remote work 2 days per week (After 90 Days Onboarding) • Travel Required: 0%
Cognizant is seeking an Azure Databricks Tech Lead to design and implement data solutions for insurance workflows in Hartford, CT. The role involves developing data pipelines and guiding junior engineers while leveraging Azure tools, Python, and SQL.
The On-site Azure Data Analytics Engineer at EDI Staffing will develop and optimize cloud-based Business Intelligence solutions, focusing on data integration and analytics. This role requires deep technical expertise in Azure technologies and collaboration with cross-functional teams.
We are looking for an experienced Azure Data Engineer to design and implement cloud-based data solutions in Phoenix, AZ. The ideal candidate will have strong technical skills in Azure services and data engineering.
SRI Tech Solutions is looking for a Senior Azure Databricks Architect to optimize data pipelines and solutions on Azure. The role requires extensive experience in data engineering, particularly with Azure Databricks, PySpark, and Python.
Sira Consulting is seeking an experienced Azure Databricks Engineer with expertise in PySpark to design and optimize big data solutions on Azure. This full-time position is based in Branchville, NJ.
Rocket is seeking a Senior Data Engineer with expertise in Microsoft Fabric and Azure Synapse to support a high-impact data integration initiative for the Finance team. This role requires hands-on development and collaboration in a multi-tenant Azure environment, based in Irving, TX.
Cognizant is seeking an Azure Databricks Tech Lead to design and implement data solutions for insurance workflows in Hartford, CT. The role involves developing data pipelines and guiding junior engineers while leveraging Azure tools, Python, and SQL.
The On-site Azure Data Analytics Engineer at EDI Staffing will develop and optimize cloud-based Business Intelligence solutions, focusing on data integration and analytics. This role requires deep technical expertise in Azure technologies and collaboration with cross-functional teams.
We are looking for an experienced Azure Data Engineer to design and implement cloud-based data solutions in Phoenix, AZ. The ideal candidate will have strong technical skills in Azure services and data engineering.
SRI Tech Solutions is looking for a Senior Azure Databricks Architect to optimize data pipelines and solutions on Azure. The role requires extensive experience in data engineering, particularly with Azure Databricks, PySpark, and Python.
Sira Consulting is seeking an experienced Azure Databricks Engineer with expertise in PySpark to design and optimize big data solutions on Azure. This full-time position is based in Branchville, NJ.
Rocket is seeking a Senior Data Engineer with expertise in Microsoft Fabric and Azure Synapse to support a high-impact data integration initiative for the Finance team. This role requires hands-on development and collaboration in a multi-tenant Azure environment, based in Irving, TX.
Cognizant is seeking an Azure Databricks Tech Lead to design and implement data solutions for insurance workflows in Hartford, CT. The role involves developing data pipelines and guiding junior engineers while leveraging Azure tools, Python, and SQL.
The On-site Azure Data Analytics Engineer at EDI Staffing will develop and optimize cloud-based Business Intelligence solutions, focusing on data integration and analytics. This role requires deep technical expertise in Azure technologies and collaboration with cross-functional teams.
The On-site Azure Data Analytics Engineer at EDI Staffing will develop and optimize cloud-based Business Intelligence solutions, focusing on data integration and analytics. This role requires deep technical expertise in Azure technologies and collaboration with cross-functional teams.