Sightline Data Engineer
Johns Hopkins Sightline: Unleash Your Impact
The Johns Hopkins Sightline Business Modernization Project is more than just a job - it's a mission. We're tearing down walls to unleash the full potential of our people. Baltimore, ranked a top 2024 destination by the New York Times, is your backdrop as you join a dynamic and entrepreneurial team to reimagine the way Johns Hopkins works. We're not just simplifying processes and transforming technology - we're empowering people to focus on what matters most: research, teaching, patient care and community impact.
Does this sound like you?
- Thrive in fast-paced environments with daily new challenges and opportunities.
- Are passionate about simplifying complex systems to change people’s lives.
- Believe in transformation that empowers, not hinders, the front-line people.
- Have experience in process improvement, technology implementation, or project management.
- Enjoy analytical thinking and problem solving.
- Communicate clearly with a team.
- Possess a desire to grow and learn.
Joining Sightline means you’ll
- Work with cutting-edge technology.
- Join a collaborative team of experts in higher education and healthcare.
- Build new skills with a team that encourages creativity, innovation, and personal growth.
- Gain valuable experience and expertise in process improvement, technology implementation, and change management.
- Enjoy a competitive salary and benefits package that includes health insurance, retirement savings plans, telework options, and generous paid time off.
Become an architect of change and help shape the future of Johns Hopkins University and Medicine alongside a passionate team in Baltimore's vibrant hub.
Sightline is seeking a highly motivated and experienced Sightline Data Engineer to join our Sightline team and lead the architecture and development of a cutting-edge Lakehouse solution in Microsoft Fabric.
Specific Duties & Responsibilities
Lakehouse Architecture Design
- Have a deep understanding of Microsoft Fabric's unified analytics engine, strong experience with cloud-native data architectures, and demonstrated expertise in modern data engineering practices for scalable and efficient data systems.
- Architect, design, and implement a modern Lakehouse leveraging Microsoft Fabric's capabilities.
- Defines and establishes best practices for ingesting, processing, and storing historical SAP data and new data streams from Workday ERP.
- Creates scalable and flexible data pipelines for hybrid data sources using native Microsoft Fabric tools such as Data Factory, Synapse Analytics, and Azure Data Lake.
SAP Historical Data Solution
- Develops a robust solution to seamlessly ingest, transform, and store historical data from SAP systems.
- Ensures data lineage, data accuracy, and compliance with regulatory requirements when working with SAP data.
- Facilitates the migration of historical data to support reporting and analytical use cases.
- Designs and implements a scalable and robust data architecture that integrates historical SAP data, new data use cases from the Workday ERP system, and data sources like EMR (Electronic Medical Records) and SIS (Student Information Systems).
Integration of New Workday Data Use Cases
- Enables seamless data integration and real-time updates from the Workday ERP system into the Lakehouse architecture.
- Collaborates with stakeholders to understand new business requirements and implement corresponding data solutions.
Reporting Warehouse Design
- Designs and implements a highly optimized reporting warehouse to support analytics and robust reporting requirements.
- Implements data models that enable self-service BI and deliver insights via tools like Power BI or other visualization platforms.
- Enhances reporting workflows by engineering efficient data marts and semantic models.
- Builds a reporting warehouse that enables comprehensive and insightful reporting, empowering stakeholders to make informed decisions with ease.
Cross-System Data Integration (EMR & SIS)
- Develops a scalable strategy to integrate, harmonize, and transform data from EMR (Electronic Medical Records) and SIS (Student Information Systems) with other enterprise systems within the Lakehouse.
- Builds reusable data pipelines and processes to support current and future cross-system reporting needs.
Scalability and Maintenance
- Designs for performance, scalability, and adaptability to support the organization's growing data needs.
- Establishes monitoring, governance, and optimization procedures for data pipelines, storage, and queries.
On-call Requirements
- As needed
- The work hours will primarily be regular day shift office hours. However, there may be times when employees are required to work off hours to respond to project inquiries, adhere to project deadlines and meet the overall needs of a 24/7 health system/university
Special Skills, Knowledge, & Abilities
- Hands-on experience with Microsoft Fabric ecosystem, including Dataverse, Synapse, Azure Data Lake, Data Factory, and related tools.
- Strong understanding of SAP data structures and integration strategies for historical data solutions.
- Deep knowledge of ERP systems like Workday and approaches for integrating new data sources into enterprise architectures.
- Experience building and optimizing data pipelines, architectures, data models, and large-scale ETL workflows.
- Proficiency in querying and modeling data using T-SQL, Python, Spark, or other relevant data engineering technologies.
- Familiarity with data reporting and visualization tools like Power BI or Tableau.
- Experience working with diverse data sources, including EMR (Electronic Medical Records) and SIS (Student Information Systems).
- Strong problem-solving skills, with a proven ability to design scalable, maintainable, and robust solutions.
- Excellent communication and collaboration skills, with experience working across technical and non-technical teams.
Minimum Qualifications
- Bachelor’s Degree required.
- Five years of related work experience focused within database management and design and business requirement gathering.
- Additional education may substitute for required experience and additional related experience may substitute for required education beyond HS Diploma/Graduation Equivalent, to the extent permitted by the JHU equivalency formula.
Preferred Qualifications
- Bachelor’s or Master’s Degree in Computer Science, Data Engineering, Information Systems, or a related field.
- Five plus years of experience in data engineering, with at least 2 years of experience in designing, developing, and implementing data lakes or lakehouses.
- Experience with cloud platforms like Azure, AWS, or GCP in the context of data engineering and analytics.
- Knowledge of machine learning model deployment pipelines or predictive analytics in the context of large datasets.
- Familiarity with data governance frameworks such as GDPR, HIPAA, or other compliance standards.
- Experience with DevOps and CI/CD workflows for data pipelines, using tools like Git, Terraform, or Azure DevOps.
Classified Title: Data Engineer
Job Posting Title (Working Title): Sightline Data Engineer
Role/Level/Range: ATP/04/PG
Starting Salary Range: $99,800 - $175,000 Annually (Commensurate w/exp.)
Employee group: Full Time
Schedule: Mon-Fri 8:30am-5:00pm
FLSA Status: Exempt
Location: Remote
Department name: 60013301-Sightline
Personnel area: University Administration
This salary range does not include all components of the Sightline compensation program. This position may be eligible for a discretionary bonus. Therefore, the actual compensation paid to the selected candidate may vary slightly from the salary range stated herein. For more information, please contact the hiring department.