Enable job alerts via email!

Data Engineer (MS Fabric)- Lead Consultant

Lingaro

Snowflake (AZ)

Remote

USD 80,000 - 120,000

Full time

30+ days ago

Boost your interview chances

Create a job specific, tailored resume for higher success rate.

Job summary

Join a forward-thinking company as a Data Engineer, where you'll play a pivotal role in designing and developing data solutions that empower data-driven decision-making. This full-time position offers the chance to work 100% remotely while collaborating with top-tier engineers and experts. You'll be responsible for creating efficient data pipelines, implementing data governance practices, and optimizing data storage architectures. With a commitment to continuous learning and development, you'll have access to extensive training programs and the opportunity to earn technology certificates. If you're passionate about big data engineering and eager to make a significant impact, this role is perfect for you!

Benefits

100% remote work
Flexibility regarding working hours
Unlimited access to Udemy learning platform
Certificate training programs
Upskilling support programs
Internal promotion opportunities
Diverse and inclusive community
Well-being and health activities
Charity and environmental support opportunities

Qualifications

  • Experience in data engineering with a focus on data processing and integration.
  • Strong programming skills in Python, PySpark, and Scala for data transformations.

Responsibilities

  • Design and implement data processing systems using Microsoft Fabric and Azure Data Analytics.
  • Build data pipelines to ingest data from various sources and ensure compatibility.

Skills

Data Engineering
Python
PySpark
Scala
Spark SQL/TSQL
Data Integration
ETL Processes
Data Governance
Problem-Solving
Communication Skills

Education

Bachelor's Degree in Computer Science
Master's Degree in Information Systems

Tools

Microsoft Fabric
Databricks
Azure Tech Stack
BigQuery
Snowflake
Hadoop
Airflow
Power BI

Job description

About Data Engineering Competency Center:

Data engineering involves the development of solutions for the collection, transformation, storage and management of data to support data-driven decision making and enable efficient data analysis by end users. It focuses on the technical aspects of data processing, integration, and delivery to ensure that data is accurate, reliable, and accessible in a timely manner. It also focuses on the scalability, cost-effectiveness, security, and supportability of the solution. Data engineering encompasses multiple toolsets and architectural concepts across on-premises and cloud stacks, including but not limited to data warehousing, data lakes, lake house, data mesh, and includes extraction, ingestion, and synchronization of structured and unstructured data across the data ecosystem. It also includes processing organization and orchestration, as well as performance optimization of data processing.

Requirements:

  • A bachelor's or master's degree in Computer Science, Information Systems, or a related field is typically required.
  • Work commercial experience as a Data Engineer or a similar role.
  • Extensive knowledge of MS Fabric components: Lakehouse, OneLake, Data Pipelines, Real-Time Analytics, Data warehouse, Power BI Integration, Semantic Models, Spark Jobs, Notebooks and Realtime Analytics, Dataflow Gen1 and Gen2, KQL.
  • Integrate Fabric capabilities for seamless data flow, governance, and collaboration across teams.
  • Strong understanding of Delta Lake, Parquet, and distributed data systems.
  • Strong programming skills in Python, PySpark, Scala and Spark SQL/TSQL for data transformations.
  • Excellent knowledge on Source control / Version Control along with CICD is a plus.
  • Strong experience in implementation and management of lake House using Databricks and Azure Tech stack (ADLS Gen2, ADF, Azure SQL).
  • Proficiency in data integration techniques, ETL processes and data pipeline architectures.
  • Solid understanding of data processing techniques such as batch processing, real-time streaming, and data integration.
  • Proficiency in working with relational and non-relational databases such as MSSQL, MySQL, PostgreSQL or Cassandra. Knowledge of data warehousing concepts and technologies like Redshift, Snowflake, or BigQuery is beneficial.
  • Good knowledge of data storage architectures, including delta lakes, data warehouses, or distributed file systems.
  • Proficient in data modeling techniques and database optimization. Knowledge of query optimization, indexing, and performance tuning is necessary for efficient data retrieval and processing.
  • Understanding of data security best practices and experience implementing data governance policies. Familiarity with data privacy regulations and compliance standards is a plus.
  • Strong problem-solving abilities to identify and resolve issues related to data processing, storage, or infrastructure. Analytical mindset to analyze and interpret complex datasets for meaningful insights.
  • Experience in designing and creating integration and unit tests will be beneficial.
  • Excellent communication skills to effectively collaborate with cross-functional teams, including data scientists, analysts, and business stakeholders. Ability to convey technical concepts to non-technical stakeholders in a clear and concise manner.
  • A passion for staying updated with emerging technologies and industry trends in the field of big data engineering. Willingness to learn and adapt to new tools and techniques to enhance data processing, storage, and analysis capabilities.
  • Proficient in database management systems such as SQL (Big Query is a must), NoSQL.
  • Candidate should be able to design, configure, and manage databases to ensure optimal performance and reliability.
  • Experience with data integration tools and techniques, such as ETL and ELT. Candidate should be able to integrate data from multiple sources and transform it into a format that is suitable for analysis.

Nice to Have: Certifications are an advantage.

  • DP-600 (Fabric Analytics Engineer Associate) certification
  • DP-700 (Microsoft Certified: Fabric Data Engineer Associate)

Tasks:

  • You will be a part of the team accountable for design, model and development of whole GCP data ecosystem for one of our Client’s (Cloud Storage, Cloud Functions, BigQuery). Involvement throughout the whole process starting with the gathering, analyzing, modelling, and documenting business/technical requirements will be needed. The role will include direct contact with clients.
  • Designing and implementing data processing systems using Microsoft Fabric, Azure Data Analytics, Databricks and other distributed frameworks like Hadoop, Spark, Snowflake, Airflow, or other similar technologies. This involves writing efficient and scalable code to process, transform, and clean large volumes of structured and unstructured data.
  • Building data pipelines to ingest data from various sources such as databases, APIs, or streaming platforms. Integrating and transforming data to ensure its compatibility with the target data model or format.
  • Designing and optimizing data storage architectures, with One lake, data lakes, data warehouses, Serverless and any distributed file systems. Implementing techniques like partitioning, compression, or indexing to optimize data storage and retrieval. Identifying and resolving bottlenecks, tuning queries, and implementing caching strategies to enhance data retrieval speed and overall system efficiency.
  • Designing and implementing data models that support efficient data storage, retrieval, and analysis. Collaborating with data scientists and analysts to understand their requirements and provide them with well-structured and optimized data for analysis and modeling purposes.
  • Utilizing frameworks like Spark to perform distributed computing tasks, such as parallel processing, distributed data processing, or machine learning algorithms.
  • Implementing security measures to protect sensitive data and ensuring compliance with data privacy regulations. Establishing data governance practices to maintain data integrity, quality, and consistency.
  • Identifying and resolving issues related to data processing, storage, or infrastructure. Monitoring system performance, identifying anomalies, and conducting root cause analysis to ensure smooth and uninterrupted data operations.
  • Collaborating with cross-functional teams including data scientists, analysts, and business stakeholders to understand their requirements and provide technical solutions. Communicating complex technical concepts to non-technical stakeholders in a clear and concise manner.
  • Independence and responsibility for delivering a solution.
  • Ability to work under Agile and Scrum development methodologies.
  • Staying updated with emerging technologies, tools, and techniques in the field of big data engineering. Exploring and recommending new technologies to enhance data processing, storage, and analysis capabilities.
  • Train and mentor junior data engineers, providing guidance and knowledge transfer.
  • Stable employment: On the market since 2008, 1300+ talents currently on board in 7 global sites.
  • 100% remote.
  • Flexibility regarding working hours.
  • Full-time position.
  • Comprehensive online onboarding program with a “Buddy” from day 1.
  • Cooperation with top-tier engineers and experts.
  • Unlimited access to the Udemy learning platform from day 1.
  • Certificate training programs: Lingarians earn 500+ technology certificates yearly.
  • Upskilling support: Capability development programs, Competency Centers, knowledge sharing sessions, community webinars, 110+ training opportunities yearly.
  • Grow as we grow as a company: 76% of our managers are internal promotions.
  • A diverse, inclusive, and values-driven community.
  • Autonomy to choose the way you work: We trust your ideas.
  • Create our community together: Refer your friends to receive bonuses.
  • Activities to support your well-being and health.
  • Plenty of opportunities to donate to charities and support the environment.
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.

Similar jobs

Azure Data Engineer - OHIO Candidates ONLY

TEKsystems

Dayton

Remote

USD 100,000 - 125,000

Yesterday
Be an early applicant

Remote Engineer, Data, I

Lensa

Jackson

Remote

USD 65,000 - 85,000

Yesterday
Be an early applicant

Oracle Cloud HCM Lead Consultant, Managed Services

Alithya

Austin

Remote

USD 90,000 - 150,000

5 days ago
Be an early applicant

Azure Data Engineer

Allegis Group

Remote

USD 80,000 - 120,000

Yesterday
Be an early applicant

Senior Data Engineer

Samsara Inc.

San Francisco

Remote

USD 112,000 - 152,000

Yesterday
Be an early applicant

Senior Product Software Data Engineer

Wolters Kluwer

Remote

USD 80,000 - 120,000

Yesterday
Be an early applicant

Senior Data Engineer

United Network for Organ Sharing

Richmond

Remote

USD 90,000 - 130,000

Yesterday
Be an early applicant

Remote Engineer, Data, I

Lensa

Owings Mills

Remote

USD 65,000 - 85,000

Yesterday
Be an early applicant

Remote Engineer, Data, I

Lensa

Atlanta

Remote

USD 65,000 - 85,000

Yesterday
Be an early applicant