
微店实时计算开发工程师(J10657)
社招全职2年以上地点:杭州状态:招聘
任职要求
任职要求: 1. 本科以上,计算机相关专业;两年以上大数据研发经验 2. 熟悉主流大数据技术,如Flink,Kafka,Hadoop、Spark、Hive、Tez、Presto、Hbase等 3. Java编程背景,有丰富的进程间通信\高并发\多线程研等研发经验 …
登录查看完整任职要求
微信扫码,1秒登录
工作职责
职责描述: 1. 负责实时数据中台融合建设,参与实时离线混部项目, 负责相关系统架构设计,资源环境规划 2. 负责大规模集群管理,架构优化,资源深化 3. 负责平台服务接口设计&研发,标准&规范制定
包括英文材料
大数据+
https://www.youtube.com/watch?v=bAyrObl7TYE
https://www.youtube.com/watch?v=H4bf_uuMC-g
With all this talk of Big Data, we got Rebecca Tickle to explain just what makes data into Big Data.
Flink+
https://nightlies.apache.org/flink/flink-docs-release-2.0/docs/learn-flink/overview/
This training presents an introduction to Apache Flink that includes just enough to get you started writing scalable streaming ETL, analytics, and event-driven applications, while leaving out a lot of (ultimately important) details.
https://www.youtube.com/watch?v=WajYe9iA2Uk&list=PLa7VYi0yPIH2GTo3vRtX8w9tgNTTyYSux
Today’s businesses are increasingly software-defined, and their business processes are being automated. Whether it’s orders and shipments, or downloads and clicks, business events can always be streamed. Flink can be used to manipulate, process, and react to these streaming events as they occur.
Kafka+
https://developer.confluent.io/what-is-apache-kafka/
https://www.youtube.com/watch?v=CU44hKLMg7k
https://www.youtube.com/watch?v=j4bqyAMMb7o&list=PLa7VYi0yPIH0KbnJQcMv5N9iW8HkZHztH
In this Apache Kafka fundamentals course, we introduce you to the basic Apache Kafka elements and APIs, as well as the broader Kafka ecosystem.
Hadoop+
https://www.runoob.com/w3cnote/hadoop-tutorial.html
Hadoop 为庞大的计算机集群提供可靠的、可伸缩的应用层计算和存储支持,它允许使用简单的编程模型跨计算机群集分布式处理大型数据集,并且支持在单台计算机到几千台计算机之间进行扩展。
[英文] Hadoop Tutorial
https://www.tutorialspoint.com/hadoop/index.htm
Hadoop is an open-source framework that allows to store and process big data in a distributed environment across clusters of computers using simple programming models.
Spark+
[英文] Learning Spark Book
https://pages.databricks.com/rs/094-YMS-629/images/LearningSpark2.0.pdf
This new edition has been updated to reflect Apache Spark’s evolution through Spark 2.x and Spark 3.0, including its expanded ecosystem of built-in and external data sources, machine learning, and streaming technologies with which Spark is tightly integrated.
Hive+
[英文] Hive Tutorial
https://www.tutorialspoint.com/hive/index.htm
Hive is a data warehouse infrastructure tool to process structured data in Hadoop. It resides on top of Hadoop to summarize Big Data, and makes querying and analyzing easy.
https://www.youtube.com/watch?v=D4HqQ8-Ja9Y
Presto+
[英文] What is Presto?
https://prestodb.io/what-is-presto/
https://www.tutorialspoint.com/apache_presto/index.htm
还有更多 •••
相关职位
社招3-5年
1、主导多智能体(Multi-Agent)系统设计与开发,支持Agent间高效协作与任务分解,构建可扩展、低耦合的智能体协作架构; 2、结合物流业务场景及数据,设计并构建具备业务感知与决策支持的智能体系统; 3、与产品、算法、业务团队紧密协作,支持大模型推理优化与部署,推动智能体系统在物流场景的落地、迭代与效果验证。
更新于 2026-01-23深圳
社招5-10年
1、主导多智能体(Multi-Agent)系统设计与开发,支持Agent间高效协作与任务分解,构建可扩展、低耦合的智能体协作架构; 2、结合物流业务场景及数据,设计并构建具备业务感知与决策支持的智能体系统; 3、与产品、算法、业务团队紧密协作,支持大模型推理优化与部署,推动智能体系统在物流场景的落地、迭代与效果验证。
更新于 2026-01-23深圳
社招5-10年
1、主导多智能体(Multi-Agent)系统设计与开发,支持Agent间高效协作与任务分解,构建可扩展、低耦合的智能体协作架构; 2、结合物流业务场景及数据,设计并构建具备业务感知与决策支持的智能体系统; 3、与产品、算法、业务团队紧密协作,支持大模型推理优化与部署,推动智能体系统在物流场景的落地、迭代与效果验证。
更新于 2026-01-21深圳