Job Description
Batch – 2025/2024/2023/2022/2021/2020 and earlier
Our Futureforce University Recruiting program is dedicated to attracting, retaining and cultivating talent. Our interns and new graduates work on real projects that affect how our business runs, giving them the opportunity to make a tangible impact on the future of our company. With offices all over the world, our recruits have the chance to collaborate and connect with fellow employees on a global scale. We offer job shadowing, mentorship programs, talent development courses, and much more.
Job Category
Intern
Job Details
We are looking for an Intern Data Engineer Analyst to join one of the scrum teams supporting our Solutions team. As a key member of this group, you will have immense opportunities to work on a broad range of technologies and diverse data domains with an eye towards making this data easy to access and to build intelligent data products. You will work alongside a team of data analyst, engineers, architects and data scientists in building predictive applications and a robust data platform. Our team shares a passion for problem solving, learning and exploring new technologies and breaking boundaries.
Responsibilities include planning, design, development, testing, automation, and driving customer adoption of our application as part of our 3 annual releases. All engineers are expected to resolve production issues as a priority. Innovation is a key part of our job as we solve complex problems. Apart from that, we allocate some percentage of every release for engineer-defined innovative solutions to improve internal testing tools or external customer experience.
Responsibilities:
- Work through all stages of the model development lifecycle: defining the business problem, capturing and structuring data, crafting end-to-end data solutions from sourcing and wrangling relevant data, to building out dashboards that reveal data insights, and presenting recommendations and business impact to partners
- Design, Build, Implement and support data pipelines, both batch and real time, to optimize data retrieval from various sources and build data engineering assets to surface useful business insights.
- Be the subject matter expert for data, pipeline design, and other related big data and programming technologies.
- Proactively identify reliability & data quality problems and drive triaging and remediation process
- Partner with data producers in understanding data sources, enable data contracts and define the data model that drives analytics
- Partner with Analysts and Data Scientists on delivering reliable data that powers actionable insights
- Evaluate various technologies and platforms in open source and proprietary products. Execute proof of concept on new technology and tools to pick the best tools and solutions.
- Harness operational excellence & continuous improvement with a can do attitude
- Extensive hands-on experience with all aspects of data management, including data modelling, data mining, data warehouses, ETL, and data visualization
- Acquiring, analysing, and structuring large amounts of data to provide key business insights across disparate sources of data (Oracle, Hive/Presto, Snowflake, Hadoop, etc.)
- Design and build self-service dashboards and tools tuned to drive revenue impact
- Document processes, use cases, and requirements in a way that can be understood by technical and business partners
- Continuously improve on existing processes, automate existing analyses, define standard methodologies, and provide thought leadership around monetization initiatives
Requirements
- Upper-level candidate in B.S./M.S. in Computer Sciences or equivalent field
- Solid understanding of data structures and distributed data processing patterns
- Knowledge of data modeling techniques in distributed storage
- Knowledge of cloud data processing architectures
- Experience with programming pipelines in languages like Java, or scripting in Python
- Experience with version control systems (Github, Stash etc..) and deployment tools
- Passionate, curious, creative, self-starter and approach problem with right methodology and intelligent decisions
- Laser focus on impact, balancing effort to value, and get things done.
Preferred:
- Experience in engineering data processing pipelines at scale
- Experience with big data technologies like Hadoop, Spark, Presto, Hive, etc.
- Experience with SQL and writing sophisticated queries (Oracle, Snowflake, Hive, Presto, Spark SQL, etc.)
- Experience in data orchestration tools like Airflow
- Data visualization experience with tools like Tableau
- Salesforce products knowledge is a plus
- Experience working with Public Cloud platforms like GCP, AWS.
- Hands-on on Salesforce products knowledge and functionality a plus.
- Familiar with production debugging techniques such as thread dump analysis and GC performance tuning