metrics

What is the maximum depth of HTML documents in practice?

↘锁芯ラ 提交于 2019-12-03 14:55:06
问题 I want to allow embedding of HTML but avoid DoS due to deeply nested HTML documents that crash some browsers. I'd like to be able to accommodate 99.9% of documents, but reject those that nest too deeply. Two closely related question: What document depth limits are built into browsers? E.g. browser X fails to parse or does not build documents with depth > some limit. Are document depth statistics for documents available on the web? Is there a site with web statistics that explains that some

How do I convert between a measure of similarity and a measure of difference (distance)?

北慕城南 提交于 2019-12-03 13:59:03
问题 Is there a general way to convert between a measure of similarity and a measure of distance? Consider a similarity measure like the number of 2-grams that two strings have in common. 2-grams('beta', 'delta') = 1 2-grams('apple', 'dappled') = 4 What if I need to feed this to an optimization algorithm that expects a measure of difference, like Levenshtein distance? This is just an example...I'm looking for a general solution, if one exists. Like how to go from Levenshtein distance to a measure

kubernetes部署metrics-server metrics-server-v0.3.4 pod报错

て烟熏妆下的殇ゞ 提交于 2019-12-03 12:04:54
[root@hadoop02 ~]# kubectl logs metrics-server-v0.3.4-76db4dd54b-s4t2d -c metrics-server -n kube-system Get http://172.16.2.220:10255/stats/summary?only_cpu_and_memory=true: dial tcp 172.16.2.220:10255: connect: connection refused, unable to fully scrape metrics from source kubelet_summary:hadoop03: unable to fetch metrics from Kubelet hadoop03 (172.16.2.160): Get http://172.16.2.160:10255/stats/summary?only_cpu_and_memory=true: dial tcp 172.16.2.160:10255: connect: connection refused, unable to fully scrape metrics from source kubelet_summary:hadoop01: unable to fetch metrics from Kubelet

滴滴实时计算发展之路及平台架构实践

拈花ヽ惹草 提交于 2019-12-03 11:56:36
toc: true title: 滴滴实时计算发展之路及平台架构实践 date: 2019-08-25 tags: Flink 大数据 滴滴的核心业务是一个实时在线服务,因此具有丰富的实时数据和实时计算场景。本文将介绍滴滴实时计算发展之路以及平台架构实践。 实时计算演进 随着滴滴业务的发展,滴滴的实时计算架构也在快速演变。到目前为止大概经历了三个阶段: 业务方自建小集群阶段; 集中式大集群、平台化阶段; SQL化阶段。 下图标识了其中重要的里程碑,稍后会给出详细阐述: 在2017年以前,滴滴并没有统一的实时计算平台,而是各个业务方自建小集群。其中用到的引擎有Storm、JStorm、Spark Streaming、Samza等。业务方自建小集群模式存在如下弊端: 需要预先采购大量机器,由于单个业务独占,资源利用率通常比较低; 缺乏有效的监控报警体系; 维护难度大,需要牵涉业务方大量精力来保障集群的稳定性; 缺乏有效技术支持,且各自沉淀的东西难以共享。 为了有效解决以上问题,滴滴从2017年年初开始构建统一的实时计算集群及平台。 技术选型上,我们基于滴滴现状选择了内部用大规模数据清洗的Spark Streaming引擎,同时引入On-YARN模式,并利用YARN的多租户体系构建了认证、鉴权、资源隔离、计费等机制。 相对于离线计算,实时计算任务对于稳定性有着更高的要求

Flink 从 0 到 1 学习 —— Flink Data transformation(转换)

假装没事ソ 提交于 2019-12-03 11:55:32
toc: true title: Flink 从 0 到 1 学习 —— Flink Data transformation(转换) date: 2018-11-04 tags: Flink 大数据 流式计算 前言 在第一篇介绍 Flink 的文章 《《从0到1学习Flink》—— Apache Flink 介绍》 中就说过 Flink 程序的结构 Flink 应用程序结构就是如上图所示: 1、Source: 数据源,Flink 在流处理和批处理上的 source 大概有 4 类:基于本地集合的 source、基于文件的 source、基于网络套接字的 source、自定义的 source。自定义的 source 常见的有 Apache kafka、Amazon Kinesis Streams、RabbitMQ、Twitter Streaming API、Apache NiFi 等,当然你也可以定义自己的 source。 2、Transformation:数据转换的各种操作,有 Map / FlatMap / Filter / KeyBy / Reduce / Fold / Aggregations / Window / WindowAll / Union / Window join / Split / Select / Project 等,操作很多

SLOC for Java projects

醉酒当歌 提交于 2019-12-03 10:38:39
I neeed a free tool to count SLOC on a Java project. I only really need the following metrics: SLOC number of comment lines optionally javadoc metrics optionally sort statistics by file type (.java, .js, .css, .html, .xml, etc) Bonus: 100% Java, I don't like mix something like sloccount with cygwin netbeans plugin or preferably , maven plugin Did you consider using Sonar (which uses its own internal tool since version 1.9 , sonar-squid, instead of JavaNCSS which has some flaws and doesn't work well with Java 1.5 or 1.6 projects)? This is the kind of report you'll get: alt text http://sonar

.NET Core微服务之基于App.Metrics+InfluxDB+Grafana实现统一性能监控

孤街醉人 提交于 2019-12-03 10:19:21
一、关于App.Metrics+InfluxDB+Grafana 1.1 App.Metrics      App.Metrics是一款开源的支持.NET Core的 监控 插件,它还可以支持跑在.NET Framework上的应用程序(版本 >= 4.5.2)。官方文档地址: https://www.app-metrics.io/ 1.2 InfluxDB      InfluxDB是一款开源的分布式 时序、时间和指标 数据库,使用go语言编写,无需外部依赖。官网地址: https://portal.influxdata.com 1.3 Grafana      Grafana是一个可视化面板( Dashboard ),有着非常漂亮的图表和布局展示,功能齐全的度量仪表盘和图形编辑器,支持Graphite、zabbix、 InfluxDB 、Prometheus和OpenTSDB作为数据源。官网地址: https://grafana.com/ 二、InfluxDB的安装与配置 2.1 Linux下的安装   请参考focus-lei的文章《 .NET Core使用App.Metrics + InfluxDB + Grafana进行APM监控 》,我这里只是做demo,所以选择了在我的Windows Server虚拟机中安装,快速一点。当然

Scala source code metrics tool (lines of code, lines of comments and so on) [closed]

余生颓废 提交于 2019-12-03 10:03:17
问题 Closed. This question is off-topic. It is not currently accepting answers. Want to improve this question? Update the question so it's on-topic for Stack Overflow. Closed 4 years ago . There seems to be so many code analysis tools supporting the java language, but I am so far unable to find one that supports scala (something simple like finding LOC would be nice)? I'm working in intellij so have tried metricsReloaded and Static plugins, but they are completely ignoring the scala files. Any

Kubernetes prometheus metrics for running pods and nodes?

匿名 (未验证) 提交于 2019-12-03 08:52:47
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: I've set up prometheus to monitor kubernetes metrics by following the prometheus documentation . A lot of useful metrics now show up in prometheus. However, I can't see any metrics referencing the status of my pods or nodes. Ideally - I'd like to be able to graph the pod status (Running, Pending, CrashLoopBackOff, Error) and nodes (NodeReady, Ready). Is this metric anywhere? If not, can I add it somewhere? And how? 回答1: The regular kubernetes setup does not expose these metrics - further discussion here . However, another service can be used

INFO No non-zero metrics in the last 30s message in filebeat

匿名 (未验证) 提交于 2019-12-03 08:48:34
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: I'm new to ELK and I'm getting issues while running logstash. I ran the logatash as defined in below link https://www.elastic.co/guide/en/logstash/current/advanced-pipeline.html But when run filebeat and logstash, Its show logstash successfully runs at port 9600. In filebeat it gives like this INFO No non-zero metrics in the last 30s Logstash is not getting input from filebeat.Please help.. the filebeat .yml is filebeat.prospectors: - input_type: log paths: - /path/to/file/logstash-tutorial.log output.logstash: hosts: ["localhost:5043"] and