Cloud Data Engineering Intern
Libertyville, IL, US, 60048-3781
Libertyville, IL, US, 60048-3781
We Make Life More Rewarding and Dignified
Location: Libertyville
Department: Information Technology
Summary:
Embark on a hands-on journey in cloud data and analytics as our Summer 2026 Cloud Data Engineering Intern. In this internship based in Libertyville, IL, you will join Hollister’s Data Platforms & AI team to help build and optimize the next generation of our cloud data architecture. You’ll work with a modern tech stack – Microsoft Azure, Microsoft Fabric, SAP Datasphere, and Power BI – to integrate and transform enterprise data into actionable insights.
From day one, you will contribute to key data initiatives: connecting data from core systems (like SAP) into our Azure analytics environment, developing robust data pipelines, and creating impactful visualizations to drive business decisions. This role offers a blend of engineering and analytics: you’ll get to write data transformations and also see the “why” behind the numbers by collaborating with business teams using the dashboards you help create. Both undergraduate and master’s students who are passionate about cloud data technologies are encouraged to apply.
For this opportunity, the expectation is to be onsite a minimum of 4 days per week in the Libertyville, IL office.
Responsibilities:
- Build & Manage Data Pipelines: Design, develop, and test data pipelines for integrating data from various sources (including enterprise SAP systems) into our Azure-based data platform. You will work with tools like Azure Data Factory or Fabric’s data integration features to ingest and transform data, ensuring pipelines are reliable and efficient. For instance, you might help configure a data flow that pulls data from SAP Datasphere into a Fabric Lakehouse, enabling advanced analytics on previously siloed data.
- Data Transformation & Modeling: Assist in cleansing and transforming raw data into analytics-ready datasets. This could involve writing SQL queries or Python scripts to combine and aggregate data, and helping to define data models (e.g., star schemas) for our cloud data warehouse. You’ll learn to optimize queries and transformations for performance, especially within the Azure Synapse and Fabric environments.
- Dashboard & Report Development: Collaborate with data analysts to develop and refine Power BI dashboards and reports for various business teams. You’ll connect Power BI to cloud datasets, create data visualizations, and iteratively improve them based on user feedback. A key project might be supporting a dashboard for a department like marketing or finance, where you’ll ensure the data is accurate and visuals clearly convey insights (such as trends in sales or operational metrics).
- Support Data Architecture Initiatives: Work closely with senior engineers on cloud data architecture projects. This might include contributing to the design of our Microsoft Fabric environment – e.g., setting up OneLake storage, configuring workspaces, or evaluating new Fabric components for enterprise use – as well as exploring ways to integrate or migrate legacy data systems (like SAP BW) into the modern platform. You’ll gain exposure to architectural decision-making and learn how different components (storage, compute, integration, visualization) fit together in a cohesive data ecosystem.
- Data Quality & Governance: Help implement data quality checks and governance practices. For example, you may assist in monitoring data pipeline runs and validating the data outputs, or in setting up rules/alerts for data freshness and accuracy. You’ll also learn about our data governance standards – such as using consistent naming conventions, data cataloging, and cloud resource tagging policies. In fact, our team emphasizes practices like tagging Azure resources for cost tracking; you might analyze usage logs or tag metadata to support our FinOps (cloud financial management) efforts.
- Collaboration & Troubleshooting: Participate in daily team stand-ups and project meetings to align on tasks. Work alongside seasoned data engineers and architects, both locally and in our global offices, to troubleshoot issues and optimize solutions. If a pipeline fails or a dashboard query is slow, you’ll be involved in diagnosing the problem (perhaps checking Azure logs or tweaking a data model) and proposing fixes. This collaboration will enhance your problem-solving skills and give you experience with agile project methodologies.
- Research & Innovation: Stay curious and bring new ideas! You will be encouraged to research emerging tools or techniques in cloud data engineering. This could mean trying out a new feature in Azure or Fabric, exploring functionalities of SAP Datasphere, or even evaluating a third-party data tool. If you discover improvements – say a faster way to load data or a useful analytics function – you’ll have the chance to demo it to the team and potentially incorporate it into our processes.
- Documentation & Training: Maintain clear documentation of your work, including pipeline designs, data workflows, and user guides for dashboards. Contribute to our team knowledge base so that others can reuse your solutions. Additionally, since this role supports AI & data literacy within the organization, you might help create short “how-to” guides or lunchtime learning sessions for colleagues on using new data tools (for example, an internal tutorial on accessing a Fabric workspace or getting started with a published Power BI report). Presenting your summer project outcomes to stakeholders will be the capstone of your internship experience.
Requirements:
- Education: Current enrollment in a Bachelor’s or Master’s program in Computer Science, Data Engineering, Information Systems, or a related field. Candidates should be returning to school after the internship (expected graduation after summer 2026 or later for undergrads, or 2026/2027 for master’s students).
- Foundational Skills: A solid understanding of databases and data concepts. Familiarity with SQL is required – you should be comfortable writing SELECT queries with JOINs, and you know what terms like “primary key” or “normalization” mean. Coursework or experience in data structures or algorithms is a plus as it underpins efficient data processing.
- Programming & Scripting: Experience with at least one programming or scripting language used in data handling, preferably Python (pandas, PySpark) or SQL-based scripting. Ability to write clear, logical code to transform data (for example, using Python to parse a CSV and load it into a database). Knowledge of version control (Git) and collaborative coding practices will be useful.
- Data Tools Knowledge: Exposure to data integration or analysis tools. This could be academic experience or self-taught projects using tools like Power BI, Tableau, or Excel for data visualization, and/or tools like ETL pipelines, databases, or cloud platforms for data processing. You don’t need to be an expert, but you should have curiosity and basic proficiency in working with data tools (for instance, having built a simple dashboard or a data pipeline in a class project).
- Analytical Mindset: Strong analytical and problem-solving skills. Ability to break down complex problems (e.g., figuring out why data doesn’t align across two systems) and a methodical approach to resolving them. Comfort with handling datasets, spotting anomalies or patterns, and using critical thinking to deduce causes and solutions.
- Communication & Teamwork: Good communication skills, both written and verbal. You should be able to document your work clearly and explain technical concepts in simple terms to team members or mentors. Collaboration is key – we’re looking for a team player who asks questions, listens to feedback, and is eager to learn from others. Since this internship is on-site, being proactive in face-to-face interactions and willing to engage with colleagues is essential.
Preferred Skills (Bonus)
- Cloud Platform Experience: Hands-on experience with cloud services, especially Microsoft Azure. Perhaps you’ve tried Azure in a class or personal project (e.g., deploying a web app, or using Azure Data Studio) or have familiarity with comparable services on AWS/GCP. Specific knowledge of Azure data services (such as Azure Synapse Analytics, Azure Data Factory, or Azure Databricks) would be a strong plus, as our environment heavily uses these technologies.
- Data Engineering Projects: Prior project or internship experience in data engineering or analytics. For example, having built a data pipeline, a simple data warehouse, or an interactive dashboard as a school project demonstrates useful practical skills. Experience with data transformation tools (ETL/ELT software, SQL scripts, etc.) or big data processing (Hadoop/Spark) could be beneficial.
- Knowledge of Microsoft Fabric & Power BI: Familiarity with Microsoft’s Fabric platform or Power BI. Maybe you’ve explored Fabric’s trial or are aware of its components (like Lakehouse, Data Factory, Power BI integration) from tech blogs or coursework. Experience creating Power BI reports or using Power BI Desktop to model data will help you contribute faster in our visualization tasks. Understanding how to publish reports to the Power BI Service or manage workspaces would be a bonus.
- SAP or ERP Data Understanding: While not required, any exposure to SAP systems or enterprise resource planning data is a plus. For instance, knowing what SAP is used for in business or having seen data from systems like SAP BW/HANA can help you quickly grasp our integration projects with SAP Datasphere. If you’ve taken courses related to enterprise systems or worked with large-scale ERP data, be sure to highlight that.
- Data Modeling & Analysis: Comfort with statistical analysis or data modeling concepts. This could include familiarity with dimensional modeling (facts/dimensions in a data warehouse) or experience analyzing data sets for insights. If you have used Python libraries (like NumPy, pandas) or analytics tools for a project (e.g., regression analysis, forecasting, etc.), it shows you can bring additional analytical thinking to the role.
- Attention to Detail: A knack for detail in working with data. For example, catching that data records don’t reconcile, or ensuring that a dashboard filter works correctly. In an environment dealing with multiple data sources and complex pipelines, being detail-oriented helps maintain data integrity (a critical trait when dealing with healthcare product data and global operations data that Hollister manages).
- Enthusiasm for Learning: A demonstrated eagerness to learn new technologies and concepts. The cloud data field evolves quickly – perhaps you’ve taught yourself a new programming language or completed an online course on your own. Show us that you’re proactive about learning (e.g., familiarity with communities or resources like Stack Overflow, Microsoft Learn, etc.) so we know you’ll embrace the chance to pick up whatever technologies our projects require.
#LI-Hybrid
The anticipated base pay range for this position is $23.00 - $30.00 per hr. The actual salary varies based on applicant's education, experience and skills.
About Hollister Incorporated
Hollister Incorporated is an independent, employee-owned company that develops, manufactures and markets healthcare products worldwide. The company spearheads the advancement of innovative products for ostomy care, continence care and critical care, and also creates educational support materials for patients and healthcare professionals. Headquartered in Libertyville, Illinois, Hollister has manufacturing and distribution centers on three continents and sells in nearly 80 countries. Hollister is a wholly owned subsidiary of The Firm of John Dickinson Schneider, Inc., and is guided both by its Mission to make life more rewarding and dignified for people who use our products and services, as well as its Vision to grow and prosper as an independent, employee-owned company, and in the process, to become better human beings.
EOE Statement
All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability or protected veteran status.
Job Req ID: 35015
Job Segment:
SAP, Cloud, ERP, Database, SQL, Technology