Browse Source

first version

master
FanBin 1 month ago
parent
commit
6c7e44fe09
  1. 1
      src/main/java/com/fanbin/KafkaMessageProducer.java
  2. 4
      src/main/java/com/fanbin/MapPutTest.java
  3. 6
      src/main/java/com/fanbin/SparkFileStreamingReader.java
  4. 7
      src/main/java/com/fanbin/SparkKafkaReader.java
  5. 2
      src/main/java/com/fanbin/leetcode/ThreeNum.java

1
src/main/java/com/fanbin/KafkaMessageProducer.java

@ -6,7 +6,6 @@ import org.apache.kafka.clients.producer.ProducerRecord;
import org.apache.kafka.common.serialization.StringSerializer; import org.apache.kafka.common.serialization.StringSerializer;
import java.util.Properties; import java.util.Properties;
import java.util.concurrent.TimeUnit;
public class KafkaMessageProducer { public class KafkaMessageProducer {
private static final String TOPIC_NAME = "topic_100"; private static final String TOPIC_NAME = "topic_100";

4
src/main/java/com/fanbin/MapPutTest.java

@ -1,10 +1,6 @@
package com.fanbin; package com.fanbin;
import java.util.HashMap;
import java.util.Map;
import java.util.Random; import java.util.Random;
import java.util.concurrent.CountDownLatch;
import java.util.stream.IntStream;
public class MapPutTest { public class MapPutTest {

6
src/main/java/com/fanbin/SparkFileStreamingReader.java

@ -1,7 +1,5 @@
package com.fanbin; package com.fanbin;
import java.util.concurrent.TimeoutException;
import org.apache.spark.sql.Dataset; import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.Row; import org.apache.spark.sql.Row;
import org.apache.spark.sql.SparkSession; import org.apache.spark.sql.SparkSession;
@ -9,8 +7,8 @@ import org.apache.spark.sql.streaming.StreamingQuery;
import org.apache.spark.sql.streaming.StreamingQueryException; import org.apache.spark.sql.streaming.StreamingQueryException;
import org.apache.spark.sql.types.DataTypes; import org.apache.spark.sql.types.DataTypes;
import org.apache.spark.sql.types.StructType; import org.apache.spark.sql.types.StructType;
import org.apache.spark.sql.types.DataTypes;
import org.apache.spark.sql.types.StructType; import java.util.concurrent.TimeoutException;
public class SparkFileStreamingReader { public class SparkFileStreamingReader {
public static void main(String[] args) throws StreamingQueryException, TimeoutException { public static void main(String[] args) throws StreamingQueryException, TimeoutException {

7
src/main/java/com/fanbin/SparkKafkaReader.java

@ -1,10 +1,5 @@
package com.fanbin; package com.fanbin;
import java.util.Collections;
import java.util.HashMap;
import java.util.Map;
import java.util.concurrent.TimeoutException;
import com.codahale.metrics.MetricRegistry; import com.codahale.metrics.MetricRegistry;
import org.apache.spark.sql.Dataset; import org.apache.spark.sql.Dataset;
import org.apache.spark.sql.Row; import org.apache.spark.sql.Row;
@ -16,6 +11,8 @@ import org.apache.spark.sql.streaming.Trigger;
import org.slf4j.Logger; import org.slf4j.Logger;
import org.slf4j.LoggerFactory; import org.slf4j.LoggerFactory;
import java.util.concurrent.TimeoutException;
public class SparkKafkaReader { public class SparkKafkaReader {
private static final Logger logger = LoggerFactory.getLogger(SparkKafkaReader.class); private static final Logger logger = LoggerFactory.getLogger(SparkKafkaReader.class);
private static final MetricRegistry metrics = new MetricRegistry(); private static final MetricRegistry metrics = new MetricRegistry();

2
src/main/java/com/fanbin/leetcode/ThreeNum.java

@ -1,9 +1,7 @@
package com.fanbin.leetcode; package com.fanbin.leetcode;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.HashMap;
import java.util.List; import java.util.List;
import java.util.Map;
class ThreeNum { class ThreeNum {
public List<List<Integer>> threeSum(int[] nums) { public List<List<Integer>> threeSum(int[] nums) {

Loading…
Cancel
Save