UN Secretariat
Department of Political and Peacebuilding Affairs-Department of Peace Operations-Shared Structure
Data Engineering Intern
Organizational Context
The UN Operations and Crisis Center (UNOCC) within the UN Secretariat in New York provides global situational awareness, crisis management, and peacekeeping support. It facilitates decision-making through crisis analysis and is establishing a crisis data warehouse to improve data access and analysis. This internship supports the UNOCC pipeline and data warehouse workstream.
Job Purpose
This internship aims to support the UN Operations and Crisis Center (UNOCC) in establishing and maintaining its crisis data warehouse. The intern will contribute to developing and managing data pipelines, ensuring data synchronization, and making reliable data accessible for analysis. This role is crucial for enhancing UNOCC's capacity to provide timely and robust crisis analytics, supporting informed decision-making in critical situations. The position offers a flexible duration of 3-6 months and can be undertaken in-person, remotely, or in a hybrid format, with a desired start date of June 1, 2026.
Responsibilities
The intern will be responsible for building, testing, troubleshooting, and maintaining Python data pipelines. This includes writing Python scripts to extract data from various UN internal and external sources via web scraping, APIs, and database connections. Harmonizing disparate datasets into a unified schema using tools like Pandas, Numpy, and SQLalchemy is a key duty. The role also involves cleaning and transforming raw data (JSON, CSV, EXCEL, PDF) into tabular formats and setting up automatic synchronization between data sources and the UNOCC data warehouse using Azure services. Additionally, the intern will collaborate with analysts to understand data requirements, improve existing pipelines, and identify new data sources to enhance crisis analytics and decision-making.
Work Experience
No professional work experience is required for this internship. However, a strong academic background in a field closely related to the internship's focus is necessary. Experience with programming languages, database technologies, ETL pipeline development, business intelligence tools, version control, and cloud services is considered advantageous.
Skills
Proficiency in Python programming. Experience with data harmonization using Pandas, Numpy, and SQLalchemy. Familiarity with data cleaning and transformation. Knowledge of ETL pipeline development (Extract, Transform, Load). Experience with Azure services (Containers, Functions) is advantageous. Ability to collaborate with analysts and contribute to system design.
Required Languages
Not informed
Desired Languages
Not informed
Summary based on official posting. Please verify all details on the official website.Official Posting ↗
Explore related opportunities