第七色在线视频,2021少妇久久久久久久久久,亚洲欧洲精品成人久久av18,亚洲国产精品特色大片观看完整版,孙宇晨将参加特朗普的晚宴

為了賬號(hào)安全,請(qǐng)及時(shí)綁定郵箱和手機(jī)立即綁定

Java代碼中Kafka的使用

標(biāo)簽:
Java

kafka的客户端,建立消费者

public class KafkaClient {    public static ConsumerConnector createConsumer() {
        Properties consumerProducer = new Properties();
        consumerProducer.put("zookeeper.connect", "ip:port");
        consumerProducer.put("group.id", "group-1");
        consumerProducer.put("serializer.class", "kafka.serializer.StringEncoder");
        consumerProducer .put("auto.offset.reset", "smallest");
        ConsumerConfig config = new ConsumerConfig(consumerProducer);
        ConsumerConnector consumer = Consumer.createJavaConsumerConnector(config);        return consumer;
    }    public static class ConsumerTest extends Thread {
        AtomicBoolean start = new AtomicBoolean(true);
        CountDownLatch c;        private String topic;
        ConsumerConnector consumer;        public ConsumerTest(String topic, ConsumerConnector consumer) {            this.consumer = consumer;            this.topic = topic;
        }        public void run() {
            Map<String, Integer> topicCountMap = new HashMap<String, Integer>();
            topicCountMap.put(topic, 1);

            Map<String, List<KafkaStream<byte[], byte[]>>> createMessageStreams = consumer.createMessageStreams(topicCountMap);
            KafkaStream<byte[], byte[]> stream = createMessageStreams.get(topic).get(0);// 获取每次接收到的这个数据

            ConsumerIterator<byte[], byte[]> iterator = stream.iterator();            while (iterator.hasNext()) {
                String j = new String(iterator.next().message());
                System.out.println(j);
            }

        }

    }    public static void main(String[] args) throws InterruptedException {
        ConsumerConnector c = KafkaClient.createConsumer();
        ConsumerTest t = new ConsumerTest("topic_sname", c);
        t.start();
        Thread.sleep(1000 * 60 * 10);
        t.interrupt();
        c.shutdown();
    }
}

kafka数据收集

public class KafkaDataCollect {    static Logger LOG = Logger.getLogger(KafkaDataCollect.class);
    ConsumerConnector consumer = KafkaClient.createConsumer();    //尺度时间戳 用于限定每日更新量
    private long ruler = 0;
    KafkaDataCollect(){        //初始化尺度时间戳
        ruler = getRuler();
    }    public long getRuler(){        long current = System.currentTimeMillis();
        String date = new java.text.SimpleDateFormat("dd/MM/yyyy").format(new java.util.Date(current));
        date = date + " 00:00:00";        long rulertime = 0; 
        try {
            rulertime = new java.text.SimpleDateFormat("dd/MM/yyyy HH:mm:ss").parse(date).getTime();
        } catch (ParseException e) {            // TODO Auto-generated catch block
            e.printStackTrace();
        }        return rulertime;
    }    
    public void dataStreamIn(String topic) {
        Map<String, Integer> topicCountMap = new HashMap<String, Integer>();
        topicCountMap.put(topic, 1);

        Map<String, List<KafkaStream<byte[], byte[]>>> createMessageStreams = consumer.createMessageStreams(topicCountMap);
        KafkaStream<byte[], byte[]> stream = createMessageStreams.get(topic).get(0);// 获取每次接收到的这个数据
        int articleCount = 0;
        ConsumerIterator<byte[], byte[]> iterator = stream.iterator();
        Jedis jedis = new Jedis(ip, 6379);        try {            while (iterator.hasNext()) {
                String msgStr = "";                if((System.currentTimeMillis() - ruler) >= 24 * 60 * 60 * 1000)
                {
                    String date = new java.text.SimpleDateFormat("dd/MM/yyyy").format(new java.util.Date(ruler));
                    msgStr = date + "Kafka data stream collected "+ articleCount+ " articles.";
                    LOG.info(msgStr);
                    articleCount = 0;
                    ruler = ruler + 24 * 60 * 60 * 1000;
                }                if((System.currentTimeMillis() - ruler) >= 8 * 60 * 60 * 1000)
                {
                    SendMessage.send("17865153777", msgStr, null, null, null);
                    msgStr = "";
                }
                String j = new String(iterator.next().message());
                Gson gson = new Gson();
                JsonFromKafkaData jsonD = gson.fromJson(j, new TypeToken<JsonFromKafkaData>() {
                }.getType());                try {
                    LOG.info(j);                    if(jsonD.getSimId()!=null && !jsonD.getSimId().contains("_"))
                        jsonD.setSimId("clusterId_"+jsonD.getSimId());
                    jedis.lpush("kafka-queue", gson.toJson(jsonD));
                    articleCount++;
                } catch (Exception e) {                    // TODO: handle exception
                    LOG.error("Input newData into queue.", e);
                }
            }
        } catch (Exception e) {
            jedis.close();
            SendMessage.send("178xxxxxxxx", "Kafka data collection process stoped.", null, null, null);            // TODO: handle exception
            e.printStackTrace();
        }
    }    
    
    public static void main(String[] args) {
        KafkaDataCollect kafkaStream = new KafkaDataCollect();
        kafkaStream.dataStreamIn("topic_name");
    }

}



作者:牛粪
链接:https://www.jianshu.com/p/a564932d91aa


點(diǎn)擊查看更多內(nèi)容
TA 點(diǎn)贊

若覺(jué)得本文不錯(cuò),就分享一下吧!

評(píng)論

作者其他優(yōu)質(zhì)文章

正在加載中
  • 推薦
  • 評(píng)論
  • 收藏
  • 共同學(xué)習(xí),寫(xiě)下你的評(píng)論
感謝您的支持,我會(huì)繼續(xù)努力的~
掃碼打賞,你說(shuō)多少就多少
贊賞金額會(huì)直接到老師賬戶
支付方式
打開(kāi)微信掃一掃,即可進(jìn)行掃碼打賞哦
今天注冊(cè)有機(jī)會(huì)得

100積分直接送

付費(fèi)專(zhuān)欄免費(fèi)學(xué)

大額優(yōu)惠券免費(fèi)領(lǐng)

立即參與 放棄機(jī)會(huì)
微信客服

購(gòu)課補(bǔ)貼
聯(lián)系客服咨詢優(yōu)惠詳情

幫助反饋 APP下載

慕課網(wǎng)APP
您的移動(dòng)學(xué)習(xí)伙伴

公眾號(hào)

掃描二維碼
關(guān)注慕課網(wǎng)微信公眾號(hào)

舉報(bào)

0/150
提交
取消