【2017.08.06】

暑假在学python,看了缪雪峰老师的教程,感觉还不错。Python确实比C++要简单,但是可能是学C++的缘故,总是觉得Python写的时候别扭,比如,在for循环中,怎样判断for循环是否结束。学完 “使用dict和set”就意味着python基础教程学完了,其实跟C++差不多,但是有几个细节需要注意。我对Python的期望就是能够写爬虫,也不知道要学到什么程度。

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
可以使用Flink的`DateTimeFormatter`类来将字符串转换为`java.time.Instant`对象,然后再使用`Instant.toEpochMilli()`方法将其转换为毫秒数,最终可以通过`org.apache.flink.api.common.eventtime.TimestampAssigner`类来创建对应的timestamp。以下是示例代码: ``` import org.apache.flink.api.common.eventtime.TimestampAssigner; import org.apache.flink.api.common.eventtime.Watermark; import org.apache.flink.api.common.typeinfo.TypeInformation; import org.apache.flink.api.java.typeutils.ResultTypeQueryable; import org.apache.flink.streaming.api.functions.AssignerWithPunctuatedWatermarks; import org.apache.flink.streaming.api.functions.timestamps.AscendingTimestampExtractor; import org.apache.flink.streaming.api.windowing.time.Time; import org.apache.flink.streaming.connectors.kafka.KafkaDeserializationSchema; import org.apache.flink.streaming.connectors.kafka.KafkaSerializationSchema; import org.apache.flink.streaming.connectors.kafka.KafkaTopicPartition; import org.apache.flink.streaming.connectors.kafka.internals.KafkaTopicPartitionState; import org.apache.flink.streaming.util.serialization.KeyedSerializationSchema; import org.apache.flink.table.api.ValidationException; import org.apache.flink.table.types.logical.LogicalType; import org.apache.flink.table.types.logical.TimestampType; import org.apache.kafka.clients.consumer.ConsumerRecord; import org.apache.kafka.clients.producer.ProducerRecord; import org.apache.kafka.common.TopicPartition; import java.nio.charset.StandardCharsets; import java.time.Instant; import java.time.ZoneOffset; import java.time.format.DateTimeFormatter; import java.time.temporal.ChronoUnit; import java.util.Collections; import java.util.List; import java.util.Map; import java.util.Objects; public class FlinkTimestampAssigner implements AssignerWithPunctuatedWatermarks<String>, ResultTypeQueryable<Long>, KafkaDeserializationSchema<String>, KafkaSerializationSchema<String>, KeyedSerializationSchema<String> { private static final long serialVersionUID = 1L; private final DateTimeFormatter formatter = DateTimeFormatter.ofPattern("yyyy-MM-dd'T'HH:mm:ssXXX"); private final LogicalType[] argumentTypes; private final int[] arguments; private final TimestampType.TimestampKind timestampKind; public FlinkTimestampAssigner(LogicalType[] argumentTypes, int[] arguments, TimestampType.TimestampKind timestampKind) { this.argumentTypes = argumentTypes; this.arguments = arguments; this.timestampKind = timestampKind; } @Override public Long extractTimestamp(String element, long recordTimestamp) { Instant instant = Instant.from(formatter.parse(element)); return instant.toEpochMilli(); } @Override public Watermark checkAndGetNextWatermark(String lastElement, long extractedTimestamp) { return new Watermark(extractedTimestamp - 1); } @Override public TypeInformation<Long> getProducedType() { return TypeInformation.of(Long.class); } @Override public boolean isEndOfStream(String nextElement) { return false; } @Override public String deserialize(ConsumerRecord<byte[], byte[]> record) throws Exception { return new String(record.value(), StandardCharsets.UTF_8); } @Override public boolean isEndOfStream(String element, KafkaTopicPartition partition, long offset) { return false; } @Override public TypeInformation<String> getProducedType() { return TypeInformation.of(String.class); } @Override public ProducerRecord<byte[], byte[]> serialize(String element, Long timestamp) { return new ProducerRecord<>(getTargetTopic(element), null, timestamp, null, element.getBytes(StandardCharsets.UTF_8)); } @Override public byte[] serializeKey(String element) { return null; } @Override public byte[] serializeValue(String element) { return element.getBytes(StandardCharsets.UTF_8); } @Override public String getTargetTopic(String element) { return null; } @Override public KafkaTopicPartition getTargetTopicPartition(String element) { return null; } @Override public void open(SerializationRuntimeContext context) throws Exception { // do nothing } @Override public void setTopicPartitionState(KafkaTopicPartition partition, KafkaTopicPartitionState state) { // do nothing } @Override public KafkaTopicPartitionState getTopicPartitionState(TopicPartition partition) { return null; } @Override public List<TopicPartition> getCommittableTopicPartitions() { return Collections.emptyList(); } @Override public Map<TopicPartition, Long> getCommittableOffsets() { return Collections.emptyMap(); } @Override public void close() throws Exception { // do nothing } } ``` 在使用时,可以将此类作为`DataStream`的`assignTimestampsAndWatermarks()`方法的参数,例如: ``` DataStream<String> input = ...; DataStream<String> withTimestampsAndWatermarks = input.assignTimestampsAndWatermarks(new FlinkTimestampAssigner(new LogicalType[]{new TimestampType(3)}, new int[]{0}, TimestampType.TimestampKind.PROCTIME)); ``` 注意,在上述代码中,`TimestampType.TimestampKind`参数可以根据实际情况进行设置。另外,如果需要将`Instant`对象转换为`java.sql.Timestamp`对象,则可以使用`java.sql.Timestamp.from(instant)`方法。

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值