滴滴27届秋储-数据研发实习生(国际化)
实习兼职数据类地点:北京状态:招聘
任职要求
1、2027届在校生,本科及以上学历,计算机或相关专业,具备互联网数仓实习工作经验者优先 2、熟悉Hadoop生态,精通Hadoop、Spark、Flink,熟悉StarRocks/Clickhouse/Doris至少一种OLAP引擎,有任务调优经验者优先 3、…
登录查看完整任职要求
微信扫码,1秒登录
工作职责
1、参与滴滴国际化数据采集、ETL、建模、开发、上线全链路数据建设工作 2、参与数据仓库ETL流程的优化及解决相关技术问题。
包括英文材料
学历+
Hadoop+
https://www.runoob.com/w3cnote/hadoop-tutorial.html
Hadoop 为庞大的计算机集群提供可靠的、可伸缩的应用层计算和存储支持,它允许使用简单的编程模型跨计算机群集分布式处理大型数据集,并且支持在单台计算机到几千台计算机之间进行扩展。
[英文] Hadoop Tutorial
https://www.tutorialspoint.com/hadoop/index.htm
Hadoop is an open-source framework that allows to store and process big data in a distributed environment across clusters of computers using simple programming models.
Spark+
[英文] Learning Spark Book
https://pages.databricks.com/rs/094-YMS-629/images/LearningSpark2.0.pdf
This new edition has been updated to reflect Apache Spark’s evolution through Spark 2.x and Spark 3.0, including its expanded ecosystem of built-in and external data sources, machine learning, and streaming technologies with which Spark is tightly integrated.
Flink+
https://nightlies.apache.org/flink/flink-docs-release-2.0/docs/learn-flink/overview/
This training presents an introduction to Apache Flink that includes just enough to get you started writing scalable streaming ETL, analytics, and event-driven applications, while leaving out a lot of (ultimately important) details.
https://www.youtube.com/watch?v=WajYe9iA2Uk&list=PLa7VYi0yPIH2GTo3vRtX8w9tgNTTyYSux
Today’s businesses are increasingly software-defined, and their business processes are being automated. Whether it’s orders and shipments, or downloads and clicks, business events can always be streamed. Flink can be used to manipulate, process, and react to these streaming events as they occur.
StarRocks+
https://docs.starrocks.io/docs/quick_start/
These Quick Start guides will help you get going with a small StarRocks environment.
https://itnext.io/introduction-to-starrocks-a-new-modern-analytical-database-1db2177d26e1
Recently, I had the opportunity to explore StarRocks which is the new kid in the block when talking about massive scale databases which are able to handle petabytes of data.
ClickHouse+
[英文] Advanced Tutorial
https://clickhouse.com/docs/tutorial
Learn how to ingest and query data in ClickHouse using the New York City taxi example dataset.
https://www.youtube.com/watch?v=FtoWGT7kS-c
ClickHouse is an open-source column-oriented DBMS for online analytical processing that allows users to generate analytical reports using SQL queries in real-time.
https://www.youtube.com/watch?v=Rhe-kUyrFUE&list=PL0Z2YDlm0b3gcY5R_MUo4fT5bPqUQ66ep
Doris+
https://doris.apache.org/docs/gettingStarted/what-is-apache-doris
还有更多 •••
相关职位
实习数据类
1、参与滴滴数据中台的智能化升级与数仓建设工作,对多源数据进行整合、清洗、存储,构建面向BI分析与AI应用的高质量数据资产,满足实时、离线、湖仓一体等多样化场景需求 2、参与面向AI的数据基础设施建设,包括资产标准化、数据湖(Iceberg/Hudi)的探索与应用,驱动业务智能决策等提供高质量数据支持 3、与产品、商分、业务团队紧密协作,深入理解业务需求与智能化产品目标,提供高效、可靠的数据支持,推动数据驱动的产品改进及AI功能落地(如智能归因、个性化推荐) 4、制定和优化数据开发规范与流程,引入AI开发工具(如大模型辅助SQL生成、代码Review、自动化质检),提升团队研发效率与代码质量,建立智能化数据开发心智 5、跟踪业界最新技术动态,重点关注AI与数据技术融合趋势(如Data+AI、数据湖仓的AI增强),将成熟技术引入实际项目,持续提升数据团队的技术先进性与业务竞争力。
更新于 2026-04-06北京
实习工程-系统类
1、面向公司研发部门的提供稳定、高效、易用的RDS产品,负责架构和运维体系设计、研发工作 2、持续优化RDS系统稳定性、性能、成本、可用性、可扩展性、可维护性等 3、深入理解业务场景和需求,发掘和思考技术演进方向,交付结果。
更新于 2026-04-06北京
实习工程-后端类
1、协助负责开发和维护大数据存储与数仓相关引擎服务,包括 HDFS、HBase、Hive、Ozone 等核心组件 2、协助负责对接平台方和业务方需求,围绕开源及内部大数据系统开展二次开发、功能优化与性能调优 3、参与分析和解决用户在存储、查询、数仓等场景中的痛点问题,提供高可用、高性能的解决方案 4、开发自动化运维和管理工具,配合监控告警体系,保障系统日常稳定运行与故障快速恢复。
更新于 2026-04-06北京